[ 531.441710] env[61995]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61995) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 531.442129] env[61995]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61995) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 531.442129] env[61995]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61995) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 531.442444] env[61995]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 531.542265] env[61995]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61995) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 531.552404] env[61995]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61995) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 532.150729] env[61995]: INFO nova.virt.driver [None req-8fbe4a83-ca07-438b-b2ec-723a0f6fe8dc None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 532.220441] env[61995]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.220602] env[61995]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.220702] env[61995]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61995) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 535.316461] env[61995]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-ee2d0288-95ae-42ac-9db2-98b937680db7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.331835] env[61995]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61995) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 535.332052] env[61995]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-9b25eaef-330b-41ce-adeb-df89fbe8ca8e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.358093] env[61995]: INFO oslo_vmware.api [-] Successfully established new session; session ID is df939. [ 535.358267] env[61995]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.138s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.358739] env[61995]: INFO nova.virt.vmwareapi.driver [None req-8fbe4a83-ca07-438b-b2ec-723a0f6fe8dc None None] VMware vCenter version: 7.0.3 [ 535.362094] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaec4b3b-1b66-4da6-b3fe-66b2d6b20a02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.379891] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24f6f7d-a76d-4030-81ae-bbc563e30db2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.385619] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22e7da5-d34a-4ccc-a9f5-2f1d6b1dc537 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.392046] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baeaa10e-83d3-4dff-a7e9-ac2800cf2881 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.404615] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab63788-36a3-4a5c-854a-9fc793c8145e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.410362] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294ddf93-b54d-4022-bc41-e67630990e00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.440016] env[61995]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-c5165558-cf2f-4fb2-bab5-ce79c69d6848 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.445082] env[61995]: DEBUG nova.virt.vmwareapi.driver [None req-8fbe4a83-ca07-438b-b2ec-723a0f6fe8dc None None] Extension org.openstack.compute already exists. {{(pid=61995) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 535.447723] env[61995]: INFO nova.compute.provider_config [None req-8fbe4a83-ca07-438b-b2ec-723a0f6fe8dc None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 535.951113] env[61995]: DEBUG nova.context [None req-8fbe4a83-ca07-438b-b2ec-723a0f6fe8dc None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),4e6f1814-07d2-46cf-afaa-3fc820f4f587(cell1) {{(pid=61995) load_cells /opt/stack/nova/nova/context.py:464}} [ 535.954109] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.954341] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.954993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.955446] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Acquiring lock "4e6f1814-07d2-46cf-afaa-3fc820f4f587" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.955634] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Lock "4e6f1814-07d2-46cf-afaa-3fc820f4f587" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.956678] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Lock "4e6f1814-07d2-46cf-afaa-3fc820f4f587" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.976358] env[61995]: INFO dbcounter [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Registered counter for database nova_cell0 [ 535.984603] env[61995]: INFO dbcounter [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Registered counter for database nova_cell1 [ 535.987937] env[61995]: DEBUG oslo_db.sqlalchemy.engines [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61995) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 535.988326] env[61995]: DEBUG oslo_db.sqlalchemy.engines [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61995) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 535.993246] env[61995]: ERROR nova.db.main.api [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.993246] env[61995]: result = function(*args, **kwargs) [ 535.993246] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.993246] env[61995]: return func(*args, **kwargs) [ 535.993246] env[61995]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 535.993246] env[61995]: result = fn(*args, **kwargs) [ 535.993246] env[61995]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 535.993246] env[61995]: return f(*args, **kwargs) [ 535.993246] env[61995]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 535.993246] env[61995]: return db.service_get_minimum_version(context, binaries) [ 535.993246] env[61995]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 535.993246] env[61995]: _check_db_access() [ 535.993246] env[61995]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 535.993246] env[61995]: stacktrace = ''.join(traceback.format_stack()) [ 535.993246] env[61995]: [ 535.994081] env[61995]: ERROR nova.db.main.api [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.994081] env[61995]: result = function(*args, **kwargs) [ 535.994081] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.994081] env[61995]: return func(*args, **kwargs) [ 535.994081] env[61995]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 535.994081] env[61995]: result = fn(*args, **kwargs) [ 535.994081] env[61995]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 535.994081] env[61995]: return f(*args, **kwargs) [ 535.994081] env[61995]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 535.994081] env[61995]: return db.service_get_minimum_version(context, binaries) [ 535.994081] env[61995]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 535.994081] env[61995]: _check_db_access() [ 535.994081] env[61995]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 535.994081] env[61995]: stacktrace = ''.join(traceback.format_stack()) [ 535.994081] env[61995]: [ 535.994451] env[61995]: WARNING nova.objects.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Failed to get minimum service version for cell 4e6f1814-07d2-46cf-afaa-3fc820f4f587 [ 535.994603] env[61995]: WARNING nova.objects.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 535.995035] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Acquiring lock "singleton_lock" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.995200] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Acquired lock "singleton_lock" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.995441] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Releasing lock "singleton_lock" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.995781] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Full set of CONF: {{(pid=61995) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 535.995965] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ******************************************************************************** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 535.996124] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Configuration options gathered from: {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 535.996267] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 535.996459] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 535.996590] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ================================================================================ {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 535.996804] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] allow_resize_to_same_host = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.996975] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] arq_binding_timeout = 300 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.997120] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] backdoor_port = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.997247] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] backdoor_socket = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.997412] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] block_device_allocate_retries = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.997572] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] block_device_allocate_retries_interval = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.997738] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cert = self.pem {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.997939] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.998128] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute_monitors = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.998301] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] config_dir = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.998474] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] config_drive_format = iso9660 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.998608] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.998771] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] config_source = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.998987] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] console_host = devstack {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.999184] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] control_exchange = nova {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.999348] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cpu_allocation_ratio = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.999511] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] daemon = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.999676] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] debug = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.999832] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] default_access_ip_network_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 535.999996] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] default_availability_zone = nova {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.000170] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] default_ephemeral_format = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.000330] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] default_green_pool_size = 1000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.000566] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.000730] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] default_schedule_zone = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.000891] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] disk_allocation_ratio = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.001065] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] enable_new_services = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.001248] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] enabled_apis = ['osapi_compute'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.001413] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] enabled_ssl_apis = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.001572] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] flat_injected = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.001727] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] force_config_drive = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.001882] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] force_raw_images = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.002096] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] graceful_shutdown_timeout = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.002265] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] heal_instance_info_cache_interval = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.002481] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] host = cpu-1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.002659] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.002825] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.002988] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.003221] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.003384] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instance_build_timeout = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.003546] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instance_delete_interval = 300 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.003711] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instance_format = [instance: %(uuid)s] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.003878] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instance_name_template = instance-%08x {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.004067] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instance_usage_audit = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.004270] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instance_usage_audit_period = month {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.004442] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.004609] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.004774] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] internal_service_availability_zone = internal {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.004931] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] key = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.005139] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] live_migration_retry_count = 30 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.005316] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_color = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.005481] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_config_append = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.005647] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.005808] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_dir = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.005966] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.006109] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_options = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.006274] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_rotate_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.006443] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_rotate_interval_type = days {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.006607] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] log_rotation_type = none {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.006736] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.006862] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.007039] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.007208] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.007336] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.007495] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] long_rpc_timeout = 1800 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.007653] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] max_concurrent_builds = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.007814] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] max_concurrent_live_migrations = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.007998] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] max_concurrent_snapshots = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.008178] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] max_local_block_devices = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.008339] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] max_logfile_count = 30 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.008501] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] max_logfile_size_mb = 200 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.008661] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] maximum_instance_delete_attempts = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.008829] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] metadata_listen = 0.0.0.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.009040] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] metadata_listen_port = 8775 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.009221] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] metadata_workers = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.009385] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] migrate_max_retries = -1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.009551] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] mkisofs_cmd = genisoimage {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.009758] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.009892] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] my_ip = 10.180.1.21 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.010068] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] network_allocate_retries = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.010249] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.010415] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.010576] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] osapi_compute_listen_port = 8774 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.010740] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] osapi_compute_unique_server_name_scope = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.010914] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] osapi_compute_workers = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.011104] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] password_length = 12 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.011268] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] periodic_enable = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.011426] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] periodic_fuzzy_delay = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.011595] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] pointer_model = usbtablet {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.011763] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] preallocate_images = none {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.011949] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] publish_errors = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.012112] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] pybasedir = /opt/stack/nova {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.012275] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ram_allocation_ratio = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.012436] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] rate_limit_burst = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.012605] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] rate_limit_except_level = CRITICAL {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.012763] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] rate_limit_interval = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.012921] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] reboot_timeout = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.013093] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] reclaim_instance_interval = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.013252] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] record = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.013420] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] reimage_timeout_per_gb = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.013583] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] report_interval = 120 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.013745] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] rescue_timeout = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.013903] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] reserved_host_cpus = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.014073] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] reserved_host_disk_mb = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.014236] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] reserved_host_memory_mb = 512 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.014398] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] reserved_huge_pages = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.014556] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] resize_confirm_window = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.014718] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] resize_fs_using_block_device = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.014876] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] resume_guests_state_on_host_boot = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.015084] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.015257] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] rpc_response_timeout = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.015418] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] run_external_periodic_tasks = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.015585] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] running_deleted_instance_action = reap {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.015757] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.015943] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] running_deleted_instance_timeout = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.016125] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler_instance_sync_interval = 120 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.016298] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_down_time = 720 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.016466] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] servicegroup_driver = db {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.016622] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] shell_completion = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.016781] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] shelved_offload_time = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.016939] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] shelved_poll_interval = 3600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.017130] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] shutdown_timeout = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.017294] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] source_is_ipv6 = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.017454] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ssl_only = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.017694] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.017883] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] sync_power_state_interval = 600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.018089] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] sync_power_state_pool_size = 1000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.018271] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] syslog_log_facility = LOG_USER {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.018430] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] tempdir = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.018593] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] timeout_nbd = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.018760] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] transport_url = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.018922] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] update_resources_interval = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.019092] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] use_cow_images = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.019255] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] use_eventlog = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.019412] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] use_journal = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.019567] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] use_json = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.019724] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] use_rootwrap_daemon = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.019882] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] use_stderr = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.020047] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] use_syslog = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.020208] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vcpu_pin_set = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.020373] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plugging_is_fatal = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.020537] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plugging_timeout = 300 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.020699] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] virt_mkfs = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.020861] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] volume_usage_poll_interval = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.021057] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] watch_log_file = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.021243] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] web = /usr/share/spice-html5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 536.021429] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_concurrency.disable_process_locking = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.021713] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.021897] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.022078] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.022252] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.022420] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.022585] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.022764] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.auth_strategy = keystone {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.022931] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.compute_link_prefix = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.023123] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.023295] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.dhcp_domain = novalocal {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.023463] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.enable_instance_password = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.023624] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.glance_link_prefix = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.023789] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.023957] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.024208] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.instance_list_per_project_cells = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.024398] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.list_records_by_skipping_down_cells = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.024563] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.local_metadata_per_cell = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.024732] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.max_limit = 1000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.024900] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.metadata_cache_expiration = 15 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.025085] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.neutron_default_tenant_id = default {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.025261] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.response_validation = warn {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.025429] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.use_neutron_default_nets = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.025598] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.025777] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.025966] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.026158] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.026332] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.vendordata_dynamic_targets = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.026494] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.vendordata_jsonfile_path = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.026673] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.026864] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.backend = dogpile.cache.memcached {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.027045] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.backend_argument = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.027221] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.config_prefix = cache.oslo {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.027391] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.dead_timeout = 60.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.027556] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.debug_cache_backend = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.027719] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.enable_retry_client = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.027939] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.enable_socket_keepalive = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.028164] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.enabled = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.028345] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.enforce_fips_mode = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.028512] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.expiration_time = 600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.028675] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.hashclient_retry_attempts = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.028840] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.029013] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_dead_retry = 300 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.029175] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_password = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.029342] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.029504] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.029665] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_pool_maxsize = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.029824] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.029983] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_sasl_enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.030177] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.030344] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.030502] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.memcache_username = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.030667] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.proxies = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.030832] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.redis_db = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.030991] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.redis_password = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.031199] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.031389] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.031562] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.redis_server = localhost:6379 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.031728] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.redis_socket_timeout = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.031888] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.redis_username = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.032062] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.retry_attempts = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.032231] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.retry_delay = 0.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.032393] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.socket_keepalive_count = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.032555] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.socket_keepalive_idle = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.032714] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.socket_keepalive_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.032871] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.tls_allowed_ciphers = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.033040] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.tls_cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.033211] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.tls_certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.033374] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.tls_enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.033532] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cache.tls_keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.033700] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.033873] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.auth_type = password {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.034047] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.034250] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.034422] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.034614] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.034791] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.cross_az_attach = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.034954] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.debug = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.035130] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.endpoint_template = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.035296] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.http_retries = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.035458] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.035615] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.035802] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.os_region_name = RegionOne {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.035976] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.036151] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cinder.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.036326] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.036486] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.cpu_dedicated_set = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.036643] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.cpu_shared_set = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.036810] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.image_type_exclude_list = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.036973] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.037155] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.037354] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.037529] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.037700] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.037881] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.resource_provider_association_refresh = 300 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.038074] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.038246] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.shutdown_retry_interval = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.038429] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.038608] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] conductor.workers = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.038787] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] console.allowed_origins = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.038949] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] console.ssl_ciphers = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.039134] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] console.ssl_minimum_version = default {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.039303] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] consoleauth.enforce_session_timeout = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.039471] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] consoleauth.token_ttl = 600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.039637] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.039795] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.039989] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.040173] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.connect_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.040333] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.connect_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.040492] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.endpoint_override = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.040652] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.040809] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.040969] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.max_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.041140] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.min_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.041330] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.region_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.041503] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.retriable_status_codes = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.041664] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.service_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.041832] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.service_type = accelerator {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.041993] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.042165] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.status_code_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.042324] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.status_code_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.042480] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.042660] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.042821] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] cyborg.version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.043010] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.backend = sqlalchemy {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.043184] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.connection = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.043350] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.connection_debug = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.043518] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.connection_parameters = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.043681] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.connection_recycle_time = 3600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.043843] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.connection_trace = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.044015] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.db_inc_retry_interval = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.044180] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.db_max_retries = 20 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.044340] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.db_max_retry_interval = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.044501] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.db_retry_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.044660] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.max_overflow = 50 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.044818] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.max_pool_size = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.044977] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.max_retries = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.045158] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.045315] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.mysql_wsrep_sync_wait = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.045472] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.pool_timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.045631] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.retry_interval = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.045810] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.slave_connection = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.045989] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.sqlite_synchronous = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.046167] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] database.use_db_reconnect = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.046343] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.backend = sqlalchemy {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.046510] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.connection = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.046675] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.connection_debug = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.046841] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.connection_parameters = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.047072] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.connection_recycle_time = 3600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.047295] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.connection_trace = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.047521] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.db_inc_retry_interval = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.047769] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.db_max_retries = 20 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.048043] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.db_max_retry_interval = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.048301] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.db_retry_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.048544] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.max_overflow = 50 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.048784] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.max_pool_size = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.049019] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.max_retries = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.049251] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.049431] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.049597] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.pool_timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.049764] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.retry_interval = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.049929] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.slave_connection = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.050150] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] api_database.sqlite_synchronous = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.050335] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] devices.enabled_mdev_types = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.050514] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.050687] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.050853] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ephemeral_storage_encryption.enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.051035] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.051212] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.api_servers = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.051378] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.051541] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.051704] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.051865] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.connect_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.052039] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.connect_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.052204] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.debug = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.052371] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.default_trusted_certificate_ids = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.052533] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.enable_certificate_validation = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.052693] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.enable_rbd_download = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.052851] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.endpoint_override = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.053082] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.053261] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.053424] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.max_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.053583] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.min_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.053746] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.num_retries = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.053914] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.rbd_ceph_conf = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.054093] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.rbd_connect_timeout = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.054263] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.rbd_pool = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.054427] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.rbd_user = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.054589] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.region_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.054746] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.retriable_status_codes = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.054903] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.service_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.055082] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.service_type = image {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.055246] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.055404] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.status_code_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.055562] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.status_code_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.055720] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.055932] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.056119] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.verify_glance_signatures = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.056281] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] glance.version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.056448] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] guestfs.debug = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.056613] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] mks.enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.056980] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.057190] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] image_cache.manager_interval = 2400 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.057362] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] image_cache.precache_concurrency = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.057533] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] image_cache.remove_unused_base_images = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.057702] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.057905] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.058092] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] image_cache.subdirectory_name = _base {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.058275] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.api_max_retries = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.058440] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.api_retry_interval = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.058600] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.058762] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.auth_type = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.058943] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.059126] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.059294] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.059456] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.conductor_group = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.059614] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.connect_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.059774] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.connect_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.059932] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.endpoint_override = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.060113] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.060273] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.060430] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.max_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.060585] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.min_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.060748] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.peer_list = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.060906] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.region_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.061075] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.retriable_status_codes = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.061242] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.serial_console_state_timeout = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.061400] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.service_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.061568] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.service_type = baremetal {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.061728] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.shard = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.061890] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.062092] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.status_code_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.062271] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.status_code_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.062433] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.062614] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.062777] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ironic.version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.062959] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.063150] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] key_manager.fixed_key = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.063334] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.063496] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.barbican_api_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.063653] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.barbican_endpoint = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.063821] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.barbican_endpoint_type = public {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.063979] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.barbican_region_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.064192] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.064360] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.064524] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.064685] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.064842] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.065015] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.number_of_retries = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.065182] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.retry_delay = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.065349] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.send_service_user_token = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.065555] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.065748] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.065918] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.verify_ssl = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.066110] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican.verify_ssl_path = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.066288] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.066453] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.auth_type = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.066611] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.066769] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.066933] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.067108] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.067267] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.067429] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.067592] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] barbican_service_user.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.067753] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.approle_role_id = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.067938] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.approle_secret_id = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.068133] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.kv_mountpoint = secret {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.068297] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.kv_path = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.068460] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.kv_version = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.068617] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.namespace = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.068774] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.root_token_id = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.068982] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.ssl_ca_crt_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.069201] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.timeout = 60.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.069402] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.use_ssl = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.069604] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.069788] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.069952] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.auth_type = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.070128] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.070289] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.070451] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.070607] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.connect_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.070765] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.connect_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.070924] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.endpoint_override = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.071099] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.071289] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.071459] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.max_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.071658] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.min_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.071829] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.region_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.072018] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.retriable_status_codes = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.072187] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.service_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.072357] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.service_type = identity {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.072519] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.072708] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.status_code_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.072899] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.status_code_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.073102] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.073353] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.073551] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] keystone.version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.073761] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.connection_uri = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.073927] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.cpu_mode = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.074139] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.074327] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.cpu_models = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.074502] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.cpu_power_governor_high = performance {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.074670] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.074832] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.cpu_power_management = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.075014] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.075182] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.device_detach_attempts = 8 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.075345] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.device_detach_timeout = 20 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.075508] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.disk_cachemodes = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.075667] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.disk_prefix = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.075833] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.enabled_perf_events = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.075997] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.file_backed_memory = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.076181] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.gid_maps = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.076343] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.hw_disk_discard = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.076500] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.hw_machine_type = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.076669] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.images_rbd_ceph_conf = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.076830] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.076989] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.077174] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.images_rbd_glance_store_name = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.077345] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.images_rbd_pool = rbd {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.077513] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.images_type = default {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.077669] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.images_volume_group = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.077829] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.inject_key = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.078030] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.inject_partition = -2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.078256] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.inject_password = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.078454] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.iscsi_iface = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.078654] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.iser_use_multipath = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.078852] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.079038] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.079210] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_downtime = 500 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.079373] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.079534] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.079699] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_inbound_addr = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.079865] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.080038] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.080202] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_scheme = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.080373] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_timeout_action = abort {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.080541] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_tunnelled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.080701] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_uri = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.080867] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.live_migration_with_native_tls = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.081037] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.max_queues = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.081229] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.081483] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.081667] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.nfs_mount_options = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.082805] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.083017] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.083195] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.083365] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.083533] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.083698] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.num_pcie_ports = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.083871] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.084052] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.pmem_namespaces = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.084220] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.quobyte_client_cfg = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.084506] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.084684] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.084857] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.085037] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.085206] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rbd_secret_uuid = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.085366] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rbd_user = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.085531] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.085704] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.085900] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rescue_image_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.086080] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rescue_kernel_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.086244] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rescue_ramdisk_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.086413] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.086570] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.rx_queue_size = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.086741] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.smbfs_mount_options = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.087024] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.087204] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.snapshot_compression = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.087366] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.snapshot_image_format = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.087588] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.087753] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.sparse_logical_volumes = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.087934] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.swtpm_enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.088127] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.swtpm_group = tss {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.088299] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.swtpm_user = tss {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.088469] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.sysinfo_serial = unique {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.088627] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.tb_cache_size = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.088784] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.tx_queue_size = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.088959] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.uid_maps = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.089134] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.use_virtio_for_bridges = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.089305] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.virt_type = kvm {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.089476] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.volume_clear = zero {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.089642] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.volume_clear_size = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.089809] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.volume_use_multipath = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.089970] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.vzstorage_cache_path = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.090153] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.090324] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.090489] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.090657] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.090938] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.091134] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.vzstorage_mount_user = stack {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.091305] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.091480] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.091654] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.auth_type = password {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.091815] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.091974] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.092161] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.092319] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.connect_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.092475] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.connect_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.092644] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.default_floating_pool = public {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.092803] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.endpoint_override = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.092964] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.extension_sync_interval = 600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.093142] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.http_retries = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.093302] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.093460] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.093617] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.max_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.093785] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.093942] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.min_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.094123] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.ovs_bridge = br-int {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.094292] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.physnets = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.094459] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.region_name = RegionOne {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.094619] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.retriable_status_codes = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.094787] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.service_metadata_proxy = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.094945] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.service_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.095127] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.service_type = network {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.095291] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.095449] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.status_code_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.095606] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.status_code_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.095779] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.095972] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.096149] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] neutron.version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.096320] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] notifications.bdms_in_notifications = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.096495] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] notifications.default_level = INFO {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.096668] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] notifications.notification_format = unversioned {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.096830] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] notifications.notify_on_state_change = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.097012] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.097197] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] pci.alias = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.097367] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] pci.device_spec = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.097529] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] pci.report_in_placement = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.097699] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.097912] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.auth_type = password {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.098134] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.098304] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.098462] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.098625] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.098781] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.connect_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.098944] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.connect_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.099113] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.default_domain_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.099271] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.default_domain_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.099426] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.domain_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.099581] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.domain_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.099737] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.endpoint_override = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.099894] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.100060] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.100218] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.max_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.100373] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.min_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.100538] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.password = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.100693] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.project_domain_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.100853] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.project_domain_name = Default {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.101029] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.project_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.101203] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.project_name = service {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.101368] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.region_name = RegionOne {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.101528] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.retriable_status_codes = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.101686] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.service_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.101850] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.service_type = placement {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.102019] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.102180] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.status_code_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.102340] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.status_code_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.102496] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.system_scope = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.102650] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.102808] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.trust_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.102962] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.user_domain_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.103142] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.user_domain_name = Default {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.103303] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.user_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.103476] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.username = nova {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.103655] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.103814] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] placement.version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.103990] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.cores = 20 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.104170] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.count_usage_from_placement = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.104340] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.104515] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.injected_file_content_bytes = 10240 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.104681] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.injected_file_path_length = 255 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.104851] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.injected_files = 5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.105030] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.instances = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.105203] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.key_pairs = 100 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.105368] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.metadata_items = 128 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.105534] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.ram = 51200 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.105698] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.recheck_quota = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.105906] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.server_group_members = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.106103] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] quota.server_groups = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.106289] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.106453] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.106614] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.image_metadata_prefilter = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.106775] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.106940] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.max_attempts = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.107117] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.max_placement_results = 1000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.107282] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.107443] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.107603] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.107775] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] scheduler.workers = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.107975] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.108162] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.108343] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.108515] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.108681] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.108847] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.109022] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.109214] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.109382] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.host_subset_size = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.109550] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.109711] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.109898] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.110095] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.isolated_hosts = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.110271] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.isolated_images = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.110437] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.110598] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.110762] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.110932] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.pci_in_placement = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.111120] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.111288] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.111448] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.111606] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.111768] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.111928] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.112102] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.track_instance_changes = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.112289] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.112460] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] metrics.required = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.112624] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] metrics.weight_multiplier = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.112786] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.112948] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] metrics.weight_setting = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.113282] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.113458] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] serial_console.enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.113635] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] serial_console.port_range = 10000:20000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.113804] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.113974] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.114155] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] serial_console.serialproxy_port = 6083 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.114322] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.114494] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.auth_type = password {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.114654] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.114810] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.114968] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.115145] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.115301] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.115470] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.send_service_user_token = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.115633] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.115809] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] service_user.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.116017] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.agent_enabled = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.116197] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.116516] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.116710] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.116881] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.html5proxy_port = 6082 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.117059] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.image_compression = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.117220] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.jpeg_compression = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.117381] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.playback_compression = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.117542] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.require_secure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.117708] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.server_listen = 127.0.0.1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.117896] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.118080] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.streaming_mode = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.118247] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] spice.zlib_compression = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.118412] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] upgrade_levels.baseapi = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.118582] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] upgrade_levels.compute = auto {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.118741] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] upgrade_levels.conductor = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.118900] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] upgrade_levels.scheduler = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.119077] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.119243] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.119403] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.119558] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.119719] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.119881] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.120049] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.120213] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.120371] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vendordata_dynamic_auth.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.120543] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.api_retry_count = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.120702] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.ca_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.120871] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.121049] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.cluster_name = testcl1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.121216] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.connection_pool_size = 10 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.121374] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.console_delay_seconds = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.121539] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.datastore_regex = ^datastore.* {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.121742] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.121941] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.host_password = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.122127] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.host_port = 443 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.122299] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.host_username = administrator@vsphere.local {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.122469] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.insecure = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.122630] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.integration_bridge = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.122794] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.maximum_objects = 100 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.122952] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.pbm_default_policy = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.123130] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.pbm_enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.123293] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.pbm_wsdl_location = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.123456] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.123616] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.serial_port_proxy_uri = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.123772] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.serial_port_service_uri = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.123937] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.task_poll_interval = 0.5 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.124122] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.use_linked_clone = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.124291] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.vnc_keymap = en-us {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.124455] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.vnc_port = 5900 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.124614] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vmware.vnc_port_total = 10000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.124795] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.auth_schemes = ['none'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.124970] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.125277] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.125467] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.125635] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.novncproxy_port = 6080 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.125834] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.server_listen = 127.0.0.1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.126097] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.126290] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.vencrypt_ca_certs = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.126452] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.vencrypt_client_cert = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.126610] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vnc.vencrypt_client_key = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.126793] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.126965] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.disable_deep_image_inspection = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.127144] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.127308] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.127468] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.127627] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.disable_rootwrap = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.127788] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.enable_numa_live_migration = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.127979] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.128160] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.128324] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.128486] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.libvirt_disable_apic = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.128646] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.128808] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.128972] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.129148] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.129305] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.129462] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.129619] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.129776] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.129933] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.130112] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.130297] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.130465] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.client_socket_timeout = 900 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.130629] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.default_pool_size = 1000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.130792] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.keep_alive = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.130958] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.max_header_line = 16384 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.131135] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.131296] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.ssl_ca_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.131456] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.ssl_cert_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.131614] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.ssl_key_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.131776] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.tcp_keepidle = 600 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.131951] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.132138] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] zvm.ca_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.132304] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] zvm.cloud_connector_url = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.132586] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.132759] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] zvm.reachable_timeout = 300 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.132944] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.enforce_new_defaults = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.133324] env[61995]: WARNING oslo_config.cfg [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 536.133512] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.enforce_scope = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.133689] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.policy_default_rule = default {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.133874] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.134063] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.policy_file = policy.yaml {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.134247] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.134408] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.134566] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.134725] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.134885] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.135064] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.135243] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.135416] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.connection_string = messaging:// {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.135581] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.enabled = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.135757] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.es_doc_type = notification {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.135931] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.es_scroll_size = 10000 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.136126] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.es_scroll_time = 2m {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.136294] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.filter_error_trace = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.136463] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.hmac_keys = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.136628] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.sentinel_service_name = mymaster {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.136797] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.socket_timeout = 0.1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.136956] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.trace_requests = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.137133] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler.trace_sqlalchemy = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.137307] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler_jaeger.process_tags = {} {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.137465] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler_jaeger.service_name_prefix = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.137627] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] profiler_otlp.service_name_prefix = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.137790] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] remote_debug.host = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.137977] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] remote_debug.port = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.138181] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.138345] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.138508] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.138668] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.138829] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.138985] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.139158] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.139320] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.139480] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.139647] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.139803] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.139970] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.140152] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.140344] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.140519] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.140687] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.140851] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.141035] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.141205] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.141368] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.141534] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.141697] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.141857] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.142034] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.142201] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.142361] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.142522] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.142680] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.142845] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.143012] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.ssl = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.143188] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.143356] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.143514] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.143681] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.143846] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.144012] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.144203] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.144368] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_notifications.retry = -1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.144552] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.144722] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.144892] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.auth_section = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.145065] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.auth_type = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.145230] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.cafile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.145385] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.certfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.145543] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.collect_timing = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.145698] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.connect_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.145866] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.connect_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.146061] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.endpoint_id = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.146228] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.endpoint_override = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.146393] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.insecure = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.146550] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.keyfile = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.146704] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.max_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.146859] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.min_version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.147024] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.region_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.147189] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.retriable_status_codes = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.147347] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.service_name = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.147502] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.service_type = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.147660] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.split_loggers = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.147817] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.status_code_retries = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.148013] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.status_code_retry_delay = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.148187] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.timeout = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.148347] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.valid_interfaces = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.148502] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_limit.version = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.148666] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_reports.file_event_handler = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.148833] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.148992] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] oslo_reports.log_dir = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.149175] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.149335] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.149492] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.149656] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.149817] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.149974] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.150162] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.150347] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_ovs_privileged.group = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.150509] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.150676] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.150841] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.151008] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] vif_plug_ovs_privileged.user = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.151189] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.151368] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.151540] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.151710] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.151877] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.152061] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.152227] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.152388] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.152562] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.152730] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_ovs.isolate_vif = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.152897] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.153074] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.153245] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.153413] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.153572] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_vif_ovs.per_port_bridge = False {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.153734] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_brick.lock_path = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.153894] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.154064] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.154234] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] privsep_osbrick.capabilities = [21] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.154389] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] privsep_osbrick.group = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.154541] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] privsep_osbrick.helper_command = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.154704] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.154863] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.155031] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] privsep_osbrick.user = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.155208] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.155370] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] nova_sys_admin.group = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.155528] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] nova_sys_admin.helper_command = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.155691] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.155864] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.156052] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] nova_sys_admin.user = None {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 536.156189] env[61995]: DEBUG oslo_service.service [None req-9ebc393f-f665-433f-8557-2914d9a93379 None None] ******************************************************************************** {{(pid=61995) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 536.156662] env[61995]: INFO nova.service [-] Starting compute node (version 0.1.0) [ 536.659783] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Getting list of instances from cluster (obj){ [ 536.659783] env[61995]: value = "domain-c8" [ 536.659783] env[61995]: _type = "ClusterComputeResource" [ 536.659783] env[61995]: } {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 536.660985] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1a8d91-9db8-4385-90b2-1914cc63f138 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.669905] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Got total of 0 instances {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 536.670472] env[61995]: WARNING nova.virt.vmwareapi.driver [None req-322d00e2-e245-4292-bfa4-605429da064c None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 536.670934] env[61995]: INFO nova.virt.node [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Generated node identity 5c086f4d-bc91-4e49-9831-bed8df133c15 [ 536.671207] env[61995]: INFO nova.virt.node [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Wrote node identity 5c086f4d-bc91-4e49-9831-bed8df133c15 to /opt/stack/data/n-cpu-1/compute_id [ 537.174123] env[61995]: WARNING nova.compute.manager [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Compute nodes ['5c086f4d-bc91-4e49-9831-bed8df133c15'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 538.180697] env[61995]: INFO nova.compute.manager [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 539.186182] env[61995]: WARNING nova.compute.manager [None req-322d00e2-e245-4292-bfa4-605429da064c None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 539.186514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.186611] env[61995]: DEBUG oslo_concurrency.lockutils [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.186756] env[61995]: DEBUG oslo_concurrency.lockutils [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.186911] env[61995]: DEBUG nova.compute.resource_tracker [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 539.187828] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a76b4b-7f7c-4e0c-b3ac-b372b720bdb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.195872] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af2d692-49e8-48a0-af87-8013fe49f412 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.209987] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0cb3d1-4a7f-4a02-b0cd-9f9e2c47219a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.216535] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b4938b-405c-4392-8324-3003c8b15343 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.246411] env[61995]: DEBUG nova.compute.resource_tracker [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181688MB free_disk=67GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 539.246623] env[61995]: DEBUG oslo_concurrency.lockutils [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.246822] env[61995]: DEBUG oslo_concurrency.lockutils [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.749626] env[61995]: WARNING nova.compute.resource_tracker [None req-322d00e2-e245-4292-bfa4-605429da064c None None] No compute node record for cpu-1:5c086f4d-bc91-4e49-9831-bed8df133c15: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 5c086f4d-bc91-4e49-9831-bed8df133c15 could not be found. [ 540.253386] env[61995]: INFO nova.compute.resource_tracker [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 5c086f4d-bc91-4e49-9831-bed8df133c15 [ 541.762354] env[61995]: DEBUG nova.compute.resource_tracker [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 541.762721] env[61995]: DEBUG nova.compute.resource_tracker [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 541.911636] env[61995]: INFO nova.scheduler.client.report [None req-322d00e2-e245-4292-bfa4-605429da064c None None] [req-2970333c-fbe1-4916-a18b-573c970e26f1] Created resource provider record via placement API for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 541.928581] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaabee07-7e38-43c0-8c4c-dde027b97312 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.936251] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f894e1-4880-4207-9299-d8268cacdfd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.967323] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b493d8a-e4ec-4162-9baf-fb91595e176c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.974310] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9a825e-1225-4b61-b803-f3ffbd015df9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.987160] env[61995]: DEBUG nova.compute.provider_tree [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 542.524020] env[61995]: DEBUG nova.scheduler.client.report [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 542.524020] env[61995]: DEBUG nova.compute.provider_tree [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 0 to 1 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 542.524020] env[61995]: DEBUG nova.compute.provider_tree [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 542.570556] env[61995]: DEBUG nova.compute.provider_tree [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 1 to 2 during operation: update_traits {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 543.077720] env[61995]: DEBUG nova.compute.resource_tracker [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 543.077720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.829s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.077720] env[61995]: DEBUG nova.service [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Creating RPC server for service compute {{(pid=61995) start /opt/stack/nova/nova/service.py:186}} [ 543.093224] env[61995]: DEBUG nova.service [None req-322d00e2-e245-4292-bfa4-605429da064c None None] Join ServiceGroup membership for this service compute {{(pid=61995) start /opt/stack/nova/nova/service.py:203}} [ 543.093397] env[61995]: DEBUG nova.servicegroup.drivers.db [None req-322d00e2-e245-4292-bfa4-605429da064c None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61995) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 544.094700] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 544.598682] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Getting list of instances from cluster (obj){ [ 544.598682] env[61995]: value = "domain-c8" [ 544.598682] env[61995]: _type = "ClusterComputeResource" [ 544.598682] env[61995]: } {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 544.599885] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0328fe-d426-463a-a5a8-f1914e3abccd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.608651] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Got total of 0 instances {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 544.608857] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 544.609189] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Getting list of instances from cluster (obj){ [ 544.609189] env[61995]: value = "domain-c8" [ 544.609189] env[61995]: _type = "ClusterComputeResource" [ 544.609189] env[61995]: } {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 544.610096] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1c050b-ab1a-4391-af29-f2f6e6fa2e19 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.617945] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Got total of 0 instances {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 578.327335] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "b6cb4bf3-291a-4aea-b64e-86597ad34dab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.327335] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "b6cb4bf3-291a-4aea-b64e-86597ad34dab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.830243] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.196392] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Acquiring lock "15c8668b-655a-4e75-ae6b-4b892a04487a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.196615] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Lock "15c8668b-655a-4e75-ae6b-4b892a04487a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.376321] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.376729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.378599] env[61995]: INFO nova.compute.claims [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.698637] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.228034] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.374225] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Acquiring lock "a7e67e95-3ac0-485a-94b9-59b68215ed7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.374499] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Lock "a7e67e95-3ac0-485a-94b9-59b68215ed7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.484313] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "1e690534-1a87-47e6-b433-70e0598e79f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.484551] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "1e690534-1a87-47e6-b433-70e0598e79f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.489159] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20be29c5-276e-4a3e-a65e-4c64099269b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.496830] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e504d7-6245-416b-96f2-cd4f5a937589 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.534018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47d5eb4-46fb-40b1-8d65-2bd65ffa8393 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.534316] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "becd14dd-9ad6-4c5d-86b9-babd03dc0d92" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.534514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "becd14dd-9ad6-4c5d-86b9-babd03dc0d92" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.541451] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd3491a-4054-4e36-8568-968bf55e9e64 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.557173] env[61995]: DEBUG nova.compute.provider_tree [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.876763] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.987020] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.038979] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.060172] env[61995]: DEBUG nova.scheduler.client.report [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.424154] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.513969] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.567131] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.190s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.569057] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 581.580023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.349s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.583403] env[61995]: INFO nova.compute.claims [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 581.588551] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.080614] env[61995]: DEBUG nova.compute.utils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.082154] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.082154] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 582.542531] env[61995]: DEBUG nova.policy [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '232abb0cac4b49ce8a6ca2647701aad8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b368dcb48f8e4b18bee5814674080022', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 582.587568] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 582.749938] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ed3dd4-3e66-4cb2-85ac-86e65d05161f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.759461] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7580e2e-041f-4b57-a9bb-e32d495cbd94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.800562] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11fe65e-c284-4297-9e28-b96437896ddf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.810970] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5717410-24c6-4bbb-b539-88c05da2edac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.828931] env[61995]: DEBUG nova.compute.provider_tree [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.880496] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Acquiring lock "9073413b-fc16-440b-84ad-6d8b826dc68a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.880496] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Lock "9073413b-fc16-440b-84ad-6d8b826dc68a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.336710] env[61995]: DEBUG nova.scheduler.client.report [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.386668] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 583.603368] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 583.639595] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.639595] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.639595] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.639754] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.639754] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.641159] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.641482] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.641635] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.642065] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.642295] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.642512] env[61995]: DEBUG nova.virt.hardware [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.643408] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1101af9b-dd29-4fdd-add7-bb348505c829 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.652239] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b5d8a0-7c66-4247-88d8-8af6f65f988a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.669654] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55083a7-c444-492a-9d09-8b3a647ee063 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.842650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Acquiring lock "ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.843552] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Lock "ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.845130] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.845130] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 583.849189] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.425s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.850928] env[61995]: INFO nova.compute.claims [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.903164] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Successfully created port: 796421a2-4dd4-4c51-99d1-ccd9ea9b23db {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.915860] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.350187] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.369019] env[61995]: DEBUG nova.compute.utils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 584.369019] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 584.369019] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 584.584152] env[61995]: DEBUG nova.policy [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4e40f8b95cc426dbba2358bb3811c3b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8409de14fa8a45249983824aef1dd076', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 584.873835] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 584.903504] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.019072] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f053a860-ee64-481a-b656-7088bd5e3284 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.035359] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23604891-11e8-4f8d-bb44-6f94fa8804d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.071661] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e987bcc0-2356-4bbb-9f6e-6427e27549ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.080511] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac3261d-2b00-41f6-bca8-26d8a4c24350 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.096793] env[61995]: DEBUG nova.compute.provider_tree [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.370481] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Successfully created port: 33fa9a01-a845-4b72-a99d-8cd1a7c7a063 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.602039] env[61995]: DEBUG nova.scheduler.client.report [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.884840] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 585.925570] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.925881] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.926109] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.926329] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.926502] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.926673] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.926903] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.927094] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.927305] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.927502] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.928059] env[61995]: DEBUG nova.virt.hardware [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.929106] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae88b2a-1fc3-434b-b58e-125c01fdf0db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.937572] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ec988f-9191-4a6f-b628-2fa056a2e824 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.110476] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.111603] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 586.114790] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.600s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.115622] env[61995]: INFO nova.compute.claims [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.626128] env[61995]: DEBUG nova.compute.utils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.630509] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 586.630768] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 586.775501] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Acquiring lock "853dd1f7-e8f7-4d55-b53b-933ed8861b3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.776083] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Lock "853dd1f7-e8f7-4d55-b53b-933ed8861b3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.883010] env[61995]: DEBUG nova.policy [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4ee8e14984f452c8465a887a8c8bf1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '522f72e2c3874eeb81c670117759d980', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 587.128540] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "f59d6eee-6c40-4a99-b141-626e68308270" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.128813] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "f59d6eee-6c40-4a99-b141-626e68308270" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.135989] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 587.160299] env[61995]: ERROR nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. [ 587.160299] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.160299] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.160299] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.160299] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.160299] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.160299] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.160299] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.160299] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.160299] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 587.160299] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.160299] env[61995]: ERROR nova.compute.manager raise self.value [ 587.160299] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.160299] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.160299] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.160299] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.160826] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.160826] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.160826] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. [ 587.160826] env[61995]: ERROR nova.compute.manager [ 587.160826] env[61995]: Traceback (most recent call last): [ 587.160826] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.160826] env[61995]: listener.cb(fileno) [ 587.160826] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.160826] env[61995]: result = function(*args, **kwargs) [ 587.160826] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.160826] env[61995]: return func(*args, **kwargs) [ 587.160826] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.160826] env[61995]: raise e [ 587.160826] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.160826] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 587.160826] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.160826] env[61995]: created_port_ids = self._update_ports_for_instance( [ 587.160826] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.160826] env[61995]: with excutils.save_and_reraise_exception(): [ 587.160826] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.160826] env[61995]: self.force_reraise() [ 587.160826] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.160826] env[61995]: raise self.value [ 587.160826] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.160826] env[61995]: updated_port = self._update_port( [ 587.160826] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.160826] env[61995]: _ensure_no_port_binding_failure(port) [ 587.160826] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.160826] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.161938] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. [ 587.161938] env[61995]: Removing descriptor: 14 [ 587.163140] env[61995]: ERROR nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Traceback (most recent call last): [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] yield resources [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self.driver.spawn(context, instance, image_meta, [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] vm_ref = self.build_virtual_machine(instance, [ 587.163140] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] for vif in network_info: [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return self._sync_wrapper(fn, *args, **kwargs) [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self.wait() [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self[:] = self._gt.wait() [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return self._exit_event.wait() [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.163482] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] result = hub.switch() [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return self.greenlet.switch() [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] result = function(*args, **kwargs) [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return func(*args, **kwargs) [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] raise e [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] nwinfo = self.network_api.allocate_for_instance( [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] created_port_ids = self._update_ports_for_instance( [ 587.163791] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] with excutils.save_and_reraise_exception(): [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self.force_reraise() [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] raise self.value [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] updated_port = self._update_port( [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] _ensure_no_port_binding_failure(port) [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] raise exception.PortBindingFailed(port_id=port['id']) [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] nova.exception.PortBindingFailed: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. [ 587.165316] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] [ 587.165786] env[61995]: INFO nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Terminating instance [ 587.168204] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.168358] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquired lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.168515] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 587.280729] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.300363] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0a9a92-c547-47ab-aff4-c6c8ae3c970d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.312355] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d7506b-f4cd-4e4d-9cad-3a78ea3fb2f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.355214] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56478f16-a15a-40d3-a8e8-7fb67e4313ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.364366] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57750c3-db63-488b-b900-ed5481dd4416 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.380127] env[61995]: DEBUG nova.compute.provider_tree [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.631563] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.718829] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Successfully created port: 521faecc-3dc6-4c06-8df0-f05b64e0d910 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 587.733614] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.813893] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.885906] env[61995]: DEBUG nova.scheduler.client.report [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.928133] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.149897] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 588.166487] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.176851] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 588.177168] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 588.177337] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.177965] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 588.177965] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.177965] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 588.178135] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 588.178221] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 588.178381] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 588.178555] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 588.178704] env[61995]: DEBUG nova.virt.hardware [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.179604] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2937c4-5b56-41c7-a0e9-040832d20437 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.190257] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c683fba7-4ab4-4a17-b15a-992526901ff5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.393351] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.393804] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.397089] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.808s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.399153] env[61995]: INFO nova.compute.claims [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.418419] env[61995]: ERROR nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. [ 588.418419] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.418419] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.418419] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.418419] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.418419] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.418419] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.418419] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.418419] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.418419] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 588.418419] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.418419] env[61995]: ERROR nova.compute.manager raise self.value [ 588.418419] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.418419] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.418419] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.418419] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.419334] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.419334] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.419334] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. [ 588.419334] env[61995]: ERROR nova.compute.manager [ 588.419334] env[61995]: Traceback (most recent call last): [ 588.419334] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.419334] env[61995]: listener.cb(fileno) [ 588.419334] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.419334] env[61995]: result = function(*args, **kwargs) [ 588.419334] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.419334] env[61995]: return func(*args, **kwargs) [ 588.419334] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.419334] env[61995]: raise e [ 588.419334] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.419334] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 588.419334] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.419334] env[61995]: created_port_ids = self._update_ports_for_instance( [ 588.419334] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.419334] env[61995]: with excutils.save_and_reraise_exception(): [ 588.419334] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.419334] env[61995]: self.force_reraise() [ 588.419334] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.419334] env[61995]: raise self.value [ 588.419334] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.419334] env[61995]: updated_port = self._update_port( [ 588.419334] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.419334] env[61995]: _ensure_no_port_binding_failure(port) [ 588.419334] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.419334] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.420865] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. [ 588.420865] env[61995]: Removing descriptor: 16 [ 588.420865] env[61995]: ERROR nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Traceback (most recent call last): [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] yield resources [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self.driver.spawn(context, instance, image_meta, [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.420865] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] vm_ref = self.build_virtual_machine(instance, [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] for vif in network_info: [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return self._sync_wrapper(fn, *args, **kwargs) [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self.wait() [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self[:] = self._gt.wait() [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return self._exit_event.wait() [ 588.422686] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] result = hub.switch() [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return self.greenlet.switch() [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] result = function(*args, **kwargs) [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return func(*args, **kwargs) [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] raise e [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] nwinfo = self.network_api.allocate_for_instance( [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.423073] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] created_port_ids = self._update_ports_for_instance( [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] with excutils.save_and_reraise_exception(): [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self.force_reraise() [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] raise self.value [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] updated_port = self._update_port( [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] _ensure_no_port_binding_failure(port) [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.423697] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] raise exception.PortBindingFailed(port_id=port['id']) [ 588.424021] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] nova.exception.PortBindingFailed: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. [ 588.424021] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] [ 588.424021] env[61995]: INFO nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Terminating instance [ 588.424021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Acquiring lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.424021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Acquired lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.428017] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.436355] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Releasing lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.436355] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 588.436355] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 588.436599] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f164955-31a7-40e2-8743-65692d0bbafc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.456043] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27815f80-328c-414f-8d64-b54c93d4c64f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.492232] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6cb4bf3-291a-4aea-b64e-86597ad34dab could not be found. [ 588.492385] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 588.492758] env[61995]: INFO nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Took 0.06 seconds to destroy the instance on the hypervisor. [ 588.493011] env[61995]: DEBUG oslo.service.loopingcall [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.493440] env[61995]: DEBUG nova.compute.manager [-] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.493533] env[61995]: DEBUG nova.network.neutron [-] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 588.521901] env[61995]: DEBUG nova.network.neutron [-] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.876654] env[61995]: DEBUG nova.compute.manager [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Received event network-changed-796421a2-4dd4-4c51-99d1-ccd9ea9b23db {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 588.876654] env[61995]: DEBUG nova.compute.manager [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Refreshing instance network info cache due to event network-changed-796421a2-4dd4-4c51-99d1-ccd9ea9b23db. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 588.876654] env[61995]: DEBUG oslo_concurrency.lockutils [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] Acquiring lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.876654] env[61995]: DEBUG oslo_concurrency.lockutils [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] Acquired lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.876654] env[61995]: DEBUG nova.network.neutron [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Refreshing network info cache for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 588.909732] env[61995]: DEBUG nova.compute.utils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.909732] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.909732] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 588.972838] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.026382] env[61995]: DEBUG nova.network.neutron [-] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.086716] env[61995]: DEBUG nova.policy [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3709c26007e499b8361e3c87075baa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '496222f4ed8a400b99044de6e97c439c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 589.303282] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.417065] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.436506] env[61995]: DEBUG nova.network.neutron [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.532718] env[61995]: INFO nova.compute.manager [-] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Took 1.04 seconds to deallocate network for instance. [ 589.538734] env[61995]: DEBUG nova.compute.claims [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.538993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.552020] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7286e87-d1a8-4312-a33e-b4b668fcb9b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.559388] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fb0100-cead-4015-a328-7a4bd0b10f17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.594455] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca51230e-e4ef-4fef-80af-cf9dda82c4f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.602799] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85396e00-fa59-4d62-a37c-5042102a731e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.616461] env[61995]: DEBUG nova.compute.provider_tree [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.629281] env[61995]: DEBUG nova.network.neutron [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.806167] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Releasing lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.806495] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.806870] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 589.807369] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b905e8e-7189-455e-8a24-8061b490b894 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.819662] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5f6488-a398-4121-8550-edb04229bc8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.848338] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15c8668b-655a-4e75-ae6b-4b892a04487a could not be found. [ 589.848564] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 589.848744] env[61995]: INFO nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 589.849108] env[61995]: DEBUG oslo.service.loopingcall [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.849212] env[61995]: DEBUG nova.compute.manager [-] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.849302] env[61995]: DEBUG nova.network.neutron [-] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.903948] env[61995]: DEBUG nova.network.neutron [-] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.122220] env[61995]: DEBUG nova.scheduler.client.report [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.133131] env[61995]: DEBUG oslo_concurrency.lockutils [req-2be429cc-6713-4517-9e91-0810ec68465e req-32255afb-4e70-40d1-bc1c-930d96aa1252 service nova] Releasing lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.280749] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Successfully created port: b93eab32-89d8-411f-8ffb-a08e16f210df {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.407335] env[61995]: DEBUG nova.network.neutron [-] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.430939] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.460164] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.461701] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.461701] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.461701] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.461701] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.461701] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.461931] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.461931] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.462555] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.462555] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.463163] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.464722] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7119ce9-6c96-4739-b87c-d6233f0bffa7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.473661] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b752dd-75b2-4253-9fa3-8a6331f5efed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.633568] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.237s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.634088] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.641399] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.724s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.641399] env[61995]: INFO nova.compute.claims [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.917281] env[61995]: INFO nova.compute.manager [-] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Took 1.06 seconds to deallocate network for instance. [ 590.921894] env[61995]: DEBUG nova.compute.claims [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.922468] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.149683] env[61995]: DEBUG nova.compute.utils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.154322] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.154518] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 591.274700] env[61995]: ERROR nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. [ 591.274700] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.274700] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.274700] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.274700] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.274700] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.274700] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.274700] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.274700] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.274700] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 591.274700] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.274700] env[61995]: ERROR nova.compute.manager raise self.value [ 591.274700] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.274700] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.274700] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.274700] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.275221] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.275221] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.275221] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. [ 591.275221] env[61995]: ERROR nova.compute.manager [ 591.276029] env[61995]: Traceback (most recent call last): [ 591.276029] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.276029] env[61995]: listener.cb(fileno) [ 591.276029] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.276029] env[61995]: result = function(*args, **kwargs) [ 591.276029] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.276029] env[61995]: return func(*args, **kwargs) [ 591.276029] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.276029] env[61995]: raise e [ 591.276029] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.276029] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 591.276029] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.276029] env[61995]: created_port_ids = self._update_ports_for_instance( [ 591.276029] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.276029] env[61995]: with excutils.save_and_reraise_exception(): [ 591.276029] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.276029] env[61995]: self.force_reraise() [ 591.276029] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.276029] env[61995]: raise self.value [ 591.276029] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.276029] env[61995]: updated_port = self._update_port( [ 591.276029] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.276029] env[61995]: _ensure_no_port_binding_failure(port) [ 591.276029] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.276029] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.276029] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. [ 591.276029] env[61995]: Removing descriptor: 17 [ 591.278637] env[61995]: ERROR nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Traceback (most recent call last): [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] yield resources [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self.driver.spawn(context, instance, image_meta, [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] vm_ref = self.build_virtual_machine(instance, [ 591.278637] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] for vif in network_info: [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return self._sync_wrapper(fn, *args, **kwargs) [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self.wait() [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self[:] = self._gt.wait() [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return self._exit_event.wait() [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.278969] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] result = hub.switch() [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return self.greenlet.switch() [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] result = function(*args, **kwargs) [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return func(*args, **kwargs) [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] raise e [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] nwinfo = self.network_api.allocate_for_instance( [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] created_port_ids = self._update_ports_for_instance( [ 591.279323] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] with excutils.save_and_reraise_exception(): [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self.force_reraise() [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] raise self.value [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] updated_port = self._update_port( [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] _ensure_no_port_binding_failure(port) [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] raise exception.PortBindingFailed(port_id=port['id']) [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] nova.exception.PortBindingFailed: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. [ 591.279650] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] [ 591.279981] env[61995]: INFO nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Terminating instance [ 591.284152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Acquiring lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.284152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Acquired lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.284152] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 591.309605] env[61995]: DEBUG nova.policy [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3709c26007e499b8361e3c87075baa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '496222f4ed8a400b99044de6e97c439c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 591.604553] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.605034] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.605096] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 591.605203] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 591.656671] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.850187] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.931763] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc161fe4-17fd-4ed4-bbd7-b2c1b2686b20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.941078] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381fe40e-7651-41ec-92cc-14a5f810885a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.974718] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd324214-def5-4218-a8db-3aa61fca78ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.986141] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f994ed49-d4cd-4e07-a321-636ab3a12f7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.007671] env[61995]: DEBUG nova.compute.provider_tree [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.111803] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 592.111927] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 592.111979] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 592.112292] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 592.112292] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 592.112437] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 592.112480] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Didn't find any instances for network info cache update. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 592.112697] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 592.112871] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 592.113318] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 592.113791] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 592.114249] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 592.114394] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 592.114673] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 592.114838] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 592.241738] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.422713] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Successfully created port: a8298ea3-78b5-4a6c-b651-4621f04fe498 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.505623] env[61995]: DEBUG nova.scheduler.client.report [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.619814] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.662398] env[61995]: DEBUG nova.compute.manager [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Received event network-vif-deleted-796421a2-4dd4-4c51-99d1-ccd9ea9b23db {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.662657] env[61995]: DEBUG nova.compute.manager [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Received event network-changed-33fa9a01-a845-4b72-a99d-8cd1a7c7a063 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.662747] env[61995]: DEBUG nova.compute.manager [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Refreshing instance network info cache due to event network-changed-33fa9a01-a845-4b72-a99d-8cd1a7c7a063. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 592.662949] env[61995]: DEBUG oslo_concurrency.lockutils [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] Acquiring lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.665522] env[61995]: DEBUG oslo_concurrency.lockutils [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] Acquired lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.665801] env[61995]: DEBUG nova.network.neutron [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Refreshing network info cache for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 592.675979] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.726378] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.728153] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.728384] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.728625] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.728819] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.730384] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.730384] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.730384] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.730384] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.730384] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.730580] env[61995]: DEBUG nova.virt.hardware [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.735505] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4de15d7-0710-4cde-88d4-614f7950f364 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.742452] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d067251d-89de-40bc-af15-178faf7fae7e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.751255] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Releasing lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.751255] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.751466] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 592.751988] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47fae376-de45-458a-a127-fa7ba577822b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.779651] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5177d27-8be8-490c-8b0d-d86504bf6d2d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.808598] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a7e67e95-3ac0-485a-94b9-59b68215ed7b could not be found. [ 592.808598] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 592.808598] env[61995]: INFO nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 592.808598] env[61995]: DEBUG oslo.service.loopingcall [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.808598] env[61995]: DEBUG nova.compute.manager [-] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.808598] env[61995]: DEBUG nova.network.neutron [-] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 592.879788] env[61995]: DEBUG nova.network.neutron [-] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.932594] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Acquiring lock "bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.932836] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Lock "bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.013539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.015414] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.016643] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.115s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.018548] env[61995]: INFO nova.compute.claims [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.309122] env[61995]: DEBUG nova.network.neutron [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.384313] env[61995]: DEBUG nova.network.neutron [-] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.439086] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.488817] env[61995]: DEBUG nova.network.neutron [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.523137] env[61995]: DEBUG nova.compute.utils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.530315] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.530315] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.744831] env[61995]: DEBUG nova.policy [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cd28faa86894a87b78cb2a17be1325a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d001167eaa149aba402f2a499832b5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 593.891142] env[61995]: INFO nova.compute.manager [-] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Took 1.08 seconds to deallocate network for instance. [ 593.896688] env[61995]: DEBUG nova.compute.claims [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.897621] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.970188] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.994071] env[61995]: DEBUG oslo_concurrency.lockutils [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] Releasing lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.994359] env[61995]: DEBUG nova.compute.manager [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Received event network-vif-deleted-33fa9a01-a845-4b72-a99d-8cd1a7c7a063 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.994633] env[61995]: DEBUG nova.compute.manager [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Received event network-changed-521faecc-3dc6-4c06-8df0-f05b64e0d910 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.994707] env[61995]: DEBUG nova.compute.manager [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Refreshing instance network info cache due to event network-changed-521faecc-3dc6-4c06-8df0-f05b64e0d910. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 593.994927] env[61995]: DEBUG oslo_concurrency.lockutils [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] Acquiring lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.995069] env[61995]: DEBUG oslo_concurrency.lockutils [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] Acquired lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.995218] env[61995]: DEBUG nova.network.neutron [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Refreshing network info cache for port 521faecc-3dc6-4c06-8df0-f05b64e0d910 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 594.034086] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 594.245146] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a08cf8-67f1-434f-94f4-d7d03ceffacb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.253179] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6ad74a-1e05-45ab-a661-5e0632a4f6a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.260892] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Acquiring lock "ac219077-ed13-45ee-8d47-ff5dd5abe89a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.261316] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Lock "ac219077-ed13-45ee-8d47-ff5dd5abe89a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.287804] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bddecf-cb92-405b-9c4c-25cd92600d12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.297596] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35960443-0ec8-4fb7-ae3b-ef578cfaa766 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.312053] env[61995]: DEBUG nova.compute.provider_tree [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.558315] env[61995]: DEBUG nova.network.neutron [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.749451] env[61995]: DEBUG nova.network.neutron [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.754209] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Acquiring lock "278cba4c-5bfb-42e0-b502-8f524ba1f380" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.754209] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Lock "278cba4c-5bfb-42e0-b502-8f524ba1f380" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.815312] env[61995]: DEBUG nova.scheduler.client.report [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.047523] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.076431] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.076688] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.076842] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.077748] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.077920] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.078113] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.078336] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.078515] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.078660] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.079293] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.079489] env[61995]: DEBUG nova.virt.hardware [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.080375] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2262aa3c-6472-4545-9da3-b78789d7093c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.089557] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bb41b7-b2f8-4121-a259-f1283e5c9532 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.257164] env[61995]: DEBUG oslo_concurrency.lockutils [req-39a55b2c-54a7-4303-8184-65f4a5f567f9 req-7952a1de-202a-47f0-82b4-937ed813c948 service nova] Releasing lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.323185] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.323185] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.324424] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.511s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.326516] env[61995]: INFO nova.compute.claims [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.443927] env[61995]: ERROR nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. [ 595.443927] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.443927] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.443927] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.443927] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.443927] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.443927] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.443927] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.443927] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.443927] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 595.443927] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.443927] env[61995]: ERROR nova.compute.manager raise self.value [ 595.443927] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.443927] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.443927] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.443927] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.444420] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.444420] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.444420] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. [ 595.444420] env[61995]: ERROR nova.compute.manager [ 595.444420] env[61995]: Traceback (most recent call last): [ 595.444420] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.444420] env[61995]: listener.cb(fileno) [ 595.444420] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.444420] env[61995]: result = function(*args, **kwargs) [ 595.444420] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.444420] env[61995]: return func(*args, **kwargs) [ 595.444420] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.444420] env[61995]: raise e [ 595.444420] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.444420] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 595.444420] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.444420] env[61995]: created_port_ids = self._update_ports_for_instance( [ 595.444420] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.444420] env[61995]: with excutils.save_and_reraise_exception(): [ 595.444420] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.444420] env[61995]: self.force_reraise() [ 595.444420] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.444420] env[61995]: raise self.value [ 595.444420] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.444420] env[61995]: updated_port = self._update_port( [ 595.444420] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.444420] env[61995]: _ensure_no_port_binding_failure(port) [ 595.444420] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.444420] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.445186] env[61995]: nova.exception.PortBindingFailed: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. [ 595.445186] env[61995]: Removing descriptor: 16 [ 595.445378] env[61995]: ERROR nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Traceback (most recent call last): [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] yield resources [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self.driver.spawn(context, instance, image_meta, [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] vm_ref = self.build_virtual_machine(instance, [ 595.445378] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] for vif in network_info: [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return self._sync_wrapper(fn, *args, **kwargs) [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self.wait() [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self[:] = self._gt.wait() [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return self._exit_event.wait() [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.445765] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] result = hub.switch() [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return self.greenlet.switch() [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] result = function(*args, **kwargs) [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return func(*args, **kwargs) [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] raise e [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] nwinfo = self.network_api.allocate_for_instance( [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] created_port_ids = self._update_ports_for_instance( [ 595.446143] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] with excutils.save_and_reraise_exception(): [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self.force_reraise() [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] raise self.value [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] updated_port = self._update_port( [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] _ensure_no_port_binding_failure(port) [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] raise exception.PortBindingFailed(port_id=port['id']) [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] nova.exception.PortBindingFailed: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. [ 595.446532] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] [ 595.446876] env[61995]: INFO nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Terminating instance [ 595.448141] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.448299] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.448457] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.591873] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Successfully created port: ce3de92c-90ef-4b18-894b-8b9361218867 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.831614] env[61995]: DEBUG nova.compute.utils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.839293] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.839549] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.028139] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.028444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.084983] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.145204] env[61995]: DEBUG nova.compute.manager [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Received event network-vif-deleted-521faecc-3dc6-4c06-8df0-f05b64e0d910 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 596.145421] env[61995]: DEBUG nova.compute.manager [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Received event network-changed-b93eab32-89d8-411f-8ffb-a08e16f210df {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 596.145610] env[61995]: DEBUG nova.compute.manager [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Refreshing instance network info cache due to event network-changed-b93eab32-89d8-411f-8ffb-a08e16f210df. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 596.145813] env[61995]: DEBUG oslo_concurrency.lockutils [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] Acquiring lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.198687] env[61995]: DEBUG nova.policy [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a37f5bdaf33941a0bfda6e0f30d13087', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '572e0416621f4ea3b6b03705439d8380', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 596.336625] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.580437] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceeb2164-07cb-469e-8edc-2296a1f61f73 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.589295] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c61964-ffc3-4b48-b460-4dd0200cadc1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.597696] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.633491] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f97a22a-a26b-4cf7-8bca-4c55956a0b92 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.642699] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d36a1b6-e76e-47d6-9d0c-531cc68a6d58 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.665624] env[61995]: DEBUG nova.compute.provider_tree [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.788114] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Acquiring lock "adb60754-9217-4b1b-b5cd-98553ff5bbbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.788399] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Lock "adb60754-9217-4b1b-b5cd-98553ff5bbbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.105608] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.105608] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 597.105608] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 597.105608] env[61995]: DEBUG oslo_concurrency.lockutils [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] Acquired lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.105608] env[61995]: DEBUG nova.network.neutron [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Refreshing network info cache for port b93eab32-89d8-411f-8ffb-a08e16f210df {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.106567] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-745b18ec-d213-4a2b-8afa-69cb51a20183 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.113936] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94baba8-5924-4a35-849d-5745826caf82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.140056] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1e690534-1a87-47e6-b433-70e0598e79f2 could not be found. [ 597.140519] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 597.140791] env[61995]: INFO nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 597.141150] env[61995]: DEBUG oslo.service.loopingcall [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.141477] env[61995]: DEBUG nova.compute.manager [-] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.141663] env[61995]: DEBUG nova.network.neutron [-] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 597.170408] env[61995]: DEBUG nova.scheduler.client.report [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.348428] env[61995]: DEBUG nova.network.neutron [-] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.350969] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.372876] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.373211] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.373264] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.373441] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.373582] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.373769] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.374049] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.374110] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.374244] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.374402] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.374621] env[61995]: DEBUG nova.virt.hardware [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.375721] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc83f53-9975-45fc-bcbb-353fbf669d4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.384861] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e947339-cc04-4c31-9054-dd8e0344a9b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.665917] env[61995]: DEBUG nova.network.neutron [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.676497] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.677051] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.680094] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.514s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.681772] env[61995]: INFO nova.compute.claims [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.798354] env[61995]: ERROR nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. [ 597.798354] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.798354] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.798354] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.798354] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.798354] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.798354] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.798354] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.798354] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.798354] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 597.798354] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.798354] env[61995]: ERROR nova.compute.manager raise self.value [ 597.798354] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.798354] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.798354] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.798354] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.807698] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.807698] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.807698] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. [ 597.807698] env[61995]: ERROR nova.compute.manager [ 597.807698] env[61995]: Traceback (most recent call last): [ 597.807698] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.807698] env[61995]: listener.cb(fileno) [ 597.807698] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.807698] env[61995]: result = function(*args, **kwargs) [ 597.807698] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.807698] env[61995]: return func(*args, **kwargs) [ 597.807698] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.807698] env[61995]: raise e [ 597.807698] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.807698] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 597.807698] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.807698] env[61995]: created_port_ids = self._update_ports_for_instance( [ 597.807698] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.807698] env[61995]: with excutils.save_and_reraise_exception(): [ 597.807698] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.807698] env[61995]: self.force_reraise() [ 597.807698] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.807698] env[61995]: raise self.value [ 597.807698] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.807698] env[61995]: updated_port = self._update_port( [ 597.807698] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.807698] env[61995]: _ensure_no_port_binding_failure(port) [ 597.807698] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.807698] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.808879] env[61995]: nova.exception.PortBindingFailed: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. [ 597.808879] env[61995]: Removing descriptor: 14 [ 597.808879] env[61995]: ERROR nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Traceback (most recent call last): [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] yield resources [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self.driver.spawn(context, instance, image_meta, [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.808879] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] vm_ref = self.build_virtual_machine(instance, [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] for vif in network_info: [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return self._sync_wrapper(fn, *args, **kwargs) [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self.wait() [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self[:] = self._gt.wait() [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return self._exit_event.wait() [ 597.809365] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] result = hub.switch() [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return self.greenlet.switch() [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] result = function(*args, **kwargs) [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return func(*args, **kwargs) [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] raise e [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] nwinfo = self.network_api.allocate_for_instance( [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.809788] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] created_port_ids = self._update_ports_for_instance( [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] with excutils.save_and_reraise_exception(): [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self.force_reraise() [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] raise self.value [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] updated_port = self._update_port( [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] _ensure_no_port_binding_failure(port) [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.810326] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] raise exception.PortBindingFailed(port_id=port['id']) [ 597.810675] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] nova.exception.PortBindingFailed: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. [ 597.810675] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] [ 597.810675] env[61995]: INFO nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Terminating instance [ 597.810675] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.810675] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.810675] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.853867] env[61995]: DEBUG nova.network.neutron [-] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.049328] env[61995]: DEBUG nova.network.neutron [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.191537] env[61995]: DEBUG nova.compute.utils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.200801] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.201585] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 598.204461] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Successfully created port: 2cfd262e-b05a-4c85-84ee-0c58efb14505 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.358931] env[61995]: INFO nova.compute.manager [-] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Took 1.22 seconds to deallocate network for instance. [ 598.361837] env[61995]: DEBUG nova.compute.claims [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 598.362015] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.374080] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.429149] env[61995]: DEBUG nova.policy [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f264e47cba149869f8b495087d56696', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '473c0c1b13c443c19993386db42ffe8d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 598.551487] env[61995]: DEBUG oslo_concurrency.lockutils [req-4254a78e-923b-42fe-b8dc-a88a51c97371 req-750c09c5-3479-4311-91c8-20e2224f6e06 service nova] Releasing lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.704990] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.758583] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.923539] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8452f5-c982-4352-a23d-fa7fe8987992 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.931820] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcec6484-55fc-4432-ba25-deb9ac60c555 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.967159] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d935c817-07da-4501-b28c-09143a848b65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.974812] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa93881-4242-4e46-8e6d-0d8ce6f4234e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.988805] env[61995]: DEBUG nova.compute.provider_tree [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.263050] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.263853] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.263853] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.264188] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95e42575-5d87-4f29-bb5c-6c5d598ad918 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.274753] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8790a96a-f151-4a4f-a4d2-1e5a911a4b01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.300172] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance becd14dd-9ad6-4c5d-86b9-babd03dc0d92 could not be found. [ 599.300292] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.300476] env[61995]: INFO nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.300716] env[61995]: DEBUG oslo.service.loopingcall [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.300921] env[61995]: DEBUG nova.compute.manager [-] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.301020] env[61995]: DEBUG nova.network.neutron [-] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.349155] env[61995]: DEBUG nova.network.neutron [-] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.491974] env[61995]: DEBUG nova.scheduler.client.report [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.596634] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Successfully created port: 6e29aea8-9a11-47d4-9ac9-42a9899f4c13 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.719894] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.755755] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.755964] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.756134] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.756324] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.756461] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.756607] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.756812] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.756977] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.757177] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.757342] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.757511] env[61995]: DEBUG nova.virt.hardware [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.758443] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36413891-c9ff-4c76-8fab-afbaec9f3613 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.768992] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb1d6c7-d1e7-4bb2-a660-47665595dd26 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.853047] env[61995]: DEBUG nova.network.neutron [-] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.002033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.002269] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.006684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.468s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.213959] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Acquiring lock "596463d7-04ca-4f29-8adf-c451310aaf15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.214458] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Lock "596463d7-04ca-4f29-8adf-c451310aaf15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.007s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.358974] env[61995]: INFO nova.compute.manager [-] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Took 1.06 seconds to deallocate network for instance. [ 600.359774] env[61995]: DEBUG nova.compute.claims [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.359937] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.513038] env[61995]: ERROR nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. [ 600.513038] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.513038] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.513038] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.513038] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.513038] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.513038] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.513038] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.513038] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.513038] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 600.513038] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.513038] env[61995]: ERROR nova.compute.manager raise self.value [ 600.513038] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.513038] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.513038] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.513038] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.513622] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.513622] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.513622] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. [ 600.513622] env[61995]: ERROR nova.compute.manager [ 600.513622] env[61995]: Traceback (most recent call last): [ 600.513622] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.513622] env[61995]: listener.cb(fileno) [ 600.513622] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.513622] env[61995]: result = function(*args, **kwargs) [ 600.513622] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.513622] env[61995]: return func(*args, **kwargs) [ 600.513622] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.513622] env[61995]: raise e [ 600.513622] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.513622] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 600.513622] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.513622] env[61995]: created_port_ids = self._update_ports_for_instance( [ 600.513622] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.513622] env[61995]: with excutils.save_and_reraise_exception(): [ 600.513622] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.513622] env[61995]: self.force_reraise() [ 600.513622] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.513622] env[61995]: raise self.value [ 600.513622] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.513622] env[61995]: updated_port = self._update_port( [ 600.513622] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.513622] env[61995]: _ensure_no_port_binding_failure(port) [ 600.513622] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.513622] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.514453] env[61995]: nova.exception.PortBindingFailed: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. [ 600.514453] env[61995]: Removing descriptor: 17 [ 600.514453] env[61995]: ERROR nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Traceback (most recent call last): [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] yield resources [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self.driver.spawn(context, instance, image_meta, [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.514453] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] vm_ref = self.build_virtual_machine(instance, [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] for vif in network_info: [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return self._sync_wrapper(fn, *args, **kwargs) [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self.wait() [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self[:] = self._gt.wait() [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return self._exit_event.wait() [ 600.514760] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] result = hub.switch() [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return self.greenlet.switch() [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] result = function(*args, **kwargs) [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return func(*args, **kwargs) [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] raise e [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] nwinfo = self.network_api.allocate_for_instance( [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.515182] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] created_port_ids = self._update_ports_for_instance( [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] with excutils.save_and_reraise_exception(): [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self.force_reraise() [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] raise self.value [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] updated_port = self._update_port( [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] _ensure_no_port_binding_failure(port) [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.515539] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] raise exception.PortBindingFailed(port_id=port['id']) [ 600.515895] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] nova.exception.PortBindingFailed: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. [ 600.515895] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] [ 600.515895] env[61995]: INFO nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Terminating instance [ 600.516192] env[61995]: DEBUG nova.compute.utils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.522618] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Acquiring lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.523443] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Acquired lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.523443] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.524778] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.524778] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.602622] env[61995]: DEBUG nova.policy [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54b695c82ecc4a95a34d0e67f082f8a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e7ab932afb04a369c1a639b4cba11d0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 600.760052] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e0aa4d-984e-4fa8-9781-8c6fec36c8a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.769577] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a7ad04-19b8-4d5f-a419-690044189556 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.810021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49dabf0-54b0-433c-a596-cebf8f839198 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.819106] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1097f70f-4ac0-43ff-a524-a5b6be46a0a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.833020] env[61995]: DEBUG nova.compute.provider_tree [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.026441] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.074393] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.260211] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.336455] env[61995]: DEBUG nova.scheduler.client.report [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.425242] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Successfully created port: a9b40936-87b7-45d7-a898-ccab46e8458c {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.766833] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Releasing lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.767096] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.767313] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 601.767635] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e72a1f5a-2f40-445f-abeb-362da614a068 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.778545] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67347e15-3ec0-408c-8b6c-a9ff5c4e9a33 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.802766] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9073413b-fc16-440b-84ad-6d8b826dc68a could not be found. [ 601.803055] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 601.803281] env[61995]: INFO nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 601.803655] env[61995]: DEBUG oslo.service.loopingcall [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.803880] env[61995]: DEBUG nova.compute.manager [-] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.803880] env[61995]: DEBUG nova.network.neutron [-] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.842262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.835s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.843426] env[61995]: ERROR nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Traceback (most recent call last): [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self.driver.spawn(context, instance, image_meta, [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] vm_ref = self.build_virtual_machine(instance, [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.843426] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] for vif in network_info: [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return self._sync_wrapper(fn, *args, **kwargs) [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self.wait() [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self[:] = self._gt.wait() [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return self._exit_event.wait() [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] result = hub.switch() [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.843850] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return self.greenlet.switch() [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] result = function(*args, **kwargs) [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] return func(*args, **kwargs) [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] raise e [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] nwinfo = self.network_api.allocate_for_instance( [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] created_port_ids = self._update_ports_for_instance( [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] with excutils.save_and_reraise_exception(): [ 601.844253] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] self.force_reraise() [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] raise self.value [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] updated_port = self._update_port( [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] _ensure_no_port_binding_failure(port) [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] raise exception.PortBindingFailed(port_id=port['id']) [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] nova.exception.PortBindingFailed: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. [ 601.844671] env[61995]: ERROR nova.compute.manager [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] [ 601.844963] env[61995]: DEBUG nova.compute.utils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.845889] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.923s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.854505] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Build of instance b6cb4bf3-291a-4aea-b64e-86597ad34dab was re-scheduled: Binding failed for port 796421a2-4dd4-4c51-99d1-ccd9ea9b23db, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 601.855220] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 601.855539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.855741] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquired lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.855928] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.869323] env[61995]: ERROR nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. [ 601.869323] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.869323] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.869323] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.869323] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.869323] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.869323] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.869323] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.869323] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.869323] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 601.869323] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.869323] env[61995]: ERROR nova.compute.manager raise self.value [ 601.869323] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.869323] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.869323] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.869323] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.869788] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.869788] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.869788] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. [ 601.869788] env[61995]: ERROR nova.compute.manager [ 601.869788] env[61995]: Traceback (most recent call last): [ 601.869929] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.869929] env[61995]: listener.cb(fileno) [ 601.869929] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.869929] env[61995]: result = function(*args, **kwargs) [ 601.869929] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.869929] env[61995]: return func(*args, **kwargs) [ 601.869929] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.869929] env[61995]: raise e [ 601.869929] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.869929] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 601.869929] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.869929] env[61995]: created_port_ids = self._update_ports_for_instance( [ 601.869929] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.869929] env[61995]: with excutils.save_and_reraise_exception(): [ 601.869929] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.869929] env[61995]: self.force_reraise() [ 601.869929] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.869929] env[61995]: raise self.value [ 601.869929] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.869929] env[61995]: updated_port = self._update_port( [ 601.869929] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.869929] env[61995]: _ensure_no_port_binding_failure(port) [ 601.869929] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.869929] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.869929] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. [ 601.869929] env[61995]: Removing descriptor: 16 [ 601.871826] env[61995]: ERROR nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Traceback (most recent call last): [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] yield resources [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self.driver.spawn(context, instance, image_meta, [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] vm_ref = self.build_virtual_machine(instance, [ 601.871826] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] for vif in network_info: [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return self._sync_wrapper(fn, *args, **kwargs) [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self.wait() [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self[:] = self._gt.wait() [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return self._exit_event.wait() [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.872348] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] result = hub.switch() [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return self.greenlet.switch() [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] result = function(*args, **kwargs) [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return func(*args, **kwargs) [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] raise e [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] nwinfo = self.network_api.allocate_for_instance( [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] created_port_ids = self._update_ports_for_instance( [ 601.872834] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] with excutils.save_and_reraise_exception(): [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self.force_reraise() [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] raise self.value [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] updated_port = self._update_port( [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] _ensure_no_port_binding_failure(port) [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] raise exception.PortBindingFailed(port_id=port['id']) [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] nova.exception.PortBindingFailed: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. [ 601.873265] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] [ 601.873751] env[61995]: INFO nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Terminating instance [ 601.874451] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Acquiring lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.874611] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Acquired lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.874806] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.900267] env[61995]: DEBUG nova.network.neutron [-] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.937289] env[61995]: DEBUG nova.compute.manager [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Received event network-vif-deleted-b93eab32-89d8-411f-8ffb-a08e16f210df {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 601.937651] env[61995]: DEBUG nova.compute.manager [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Received event network-changed-a8298ea3-78b5-4a6c-b651-4621f04fe498 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 601.937651] env[61995]: DEBUG nova.compute.manager [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Refreshing instance network info cache due to event network-changed-a8298ea3-78b5-4a6c-b651-4621f04fe498. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 601.937964] env[61995]: DEBUG oslo_concurrency.lockutils [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] Acquiring lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.938192] env[61995]: DEBUG oslo_concurrency.lockutils [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] Acquired lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.939083] env[61995]: DEBUG nova.network.neutron [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Refreshing network info cache for port a8298ea3-78b5-4a6c-b651-4621f04fe498 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 602.042190] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.077846] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.077846] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.078060] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.078402] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.078496] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.078661] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.079261] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.079456] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.079507] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.079647] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.079817] env[61995]: DEBUG nova.virt.hardware [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.080682] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c8fce2-155c-419b-9848-b73124d98bfb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.090972] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2287c11f-3eee-4674-947e-7b6ed13c0f72 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.405977] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.410319] env[61995]: DEBUG nova.network.neutron [-] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.419140] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.487882] env[61995]: DEBUG nova.network.neutron [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.614380] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e928f6b-292a-4889-acb5-cd236a7fd10f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.621827] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac10bce-342b-4341-8bb9-dc476fecdfc7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.626374] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.660057] env[61995]: DEBUG nova.network.neutron [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.665162] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.667403] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ab089e-b7ca-40e5-a952-7dece87ceb7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.678759] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b370f2-a3ee-4651-a792-0a7876303ca2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.694887] env[61995]: DEBUG nova.compute.provider_tree [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.851694] env[61995]: ERROR nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. [ 602.851694] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.851694] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.851694] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.851694] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.851694] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.851694] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.851694] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.851694] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.851694] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 602.851694] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.851694] env[61995]: ERROR nova.compute.manager raise self.value [ 602.851694] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.851694] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.851694] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.851694] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.852178] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.852178] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.852178] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. [ 602.852178] env[61995]: ERROR nova.compute.manager [ 602.852178] env[61995]: Traceback (most recent call last): [ 602.852178] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.852178] env[61995]: listener.cb(fileno) [ 602.852178] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.852178] env[61995]: result = function(*args, **kwargs) [ 602.852178] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.852178] env[61995]: return func(*args, **kwargs) [ 602.852178] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.852178] env[61995]: raise e [ 602.852178] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.852178] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 602.852178] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.852178] env[61995]: created_port_ids = self._update_ports_for_instance( [ 602.852178] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.852178] env[61995]: with excutils.save_and_reraise_exception(): [ 602.852178] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.852178] env[61995]: self.force_reraise() [ 602.852178] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.852178] env[61995]: raise self.value [ 602.852178] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.852178] env[61995]: updated_port = self._update_port( [ 602.852178] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.852178] env[61995]: _ensure_no_port_binding_failure(port) [ 602.852178] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.852178] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.852875] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. [ 602.852875] env[61995]: Removing descriptor: 19 [ 602.852875] env[61995]: ERROR nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Traceback (most recent call last): [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] yield resources [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self.driver.spawn(context, instance, image_meta, [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.852875] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] vm_ref = self.build_virtual_machine(instance, [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] for vif in network_info: [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return self._sync_wrapper(fn, *args, **kwargs) [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self.wait() [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self[:] = self._gt.wait() [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return self._exit_event.wait() [ 602.853187] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] result = hub.switch() [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return self.greenlet.switch() [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] result = function(*args, **kwargs) [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return func(*args, **kwargs) [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] raise e [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] nwinfo = self.network_api.allocate_for_instance( [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.853597] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] created_port_ids = self._update_ports_for_instance( [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] with excutils.save_and_reraise_exception(): [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self.force_reraise() [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] raise self.value [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] updated_port = self._update_port( [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] _ensure_no_port_binding_failure(port) [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.853952] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] raise exception.PortBindingFailed(port_id=port['id']) [ 602.854311] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] nova.exception.PortBindingFailed: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. [ 602.854311] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] [ 602.854311] env[61995]: INFO nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Terminating instance [ 602.855404] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Acquiring lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.855703] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Acquired lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.856904] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.922336] env[61995]: INFO nova.compute.manager [-] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Took 1.12 seconds to deallocate network for instance. [ 602.924848] env[61995]: DEBUG nova.compute.claims [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.925092] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.129879] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Releasing lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.130459] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.130630] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 603.130913] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fda41070-e118-4ec0-bd23-b7646e0bbcb4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.145979] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066fd0fd-728d-414d-95a5-46d04d929a26 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.166842] env[61995]: DEBUG oslo_concurrency.lockutils [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] Releasing lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.167078] env[61995]: DEBUG nova.compute.manager [req-c070d1af-cf04-4d97-a905-e23c1c8616ed req-fef6e98e-c746-45df-92c4-20e38b0b7f47 service nova] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Received event network-vif-deleted-a8298ea3-78b5-4a6c-b651-4621f04fe498 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.167696] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ebd55d0b-c807-4adc-b1e4-b738f61b9bb3 could not be found. [ 603.167912] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 603.168128] env[61995]: INFO nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.168384] env[61995]: DEBUG oslo.service.loopingcall [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.168584] env[61995]: DEBUG nova.compute.manager [-] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.168670] env[61995]: DEBUG nova.network.neutron [-] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.172088] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Releasing lock "refresh_cache-b6cb4bf3-291a-4aea-b64e-86597ad34dab" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.172320] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.172481] env[61995]: DEBUG nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.172643] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.198616] env[61995]: DEBUG nova.scheduler.client.report [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.206918] env[61995]: DEBUG nova.network.neutron [-] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.212740] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.384093] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.619624] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.707858] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.862s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.708665] env[61995]: ERROR nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Traceback (most recent call last): [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self.driver.spawn(context, instance, image_meta, [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] vm_ref = self.build_virtual_machine(instance, [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.708665] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] for vif in network_info: [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return self._sync_wrapper(fn, *args, **kwargs) [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self.wait() [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self[:] = self._gt.wait() [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return self._exit_event.wait() [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] result = hub.switch() [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.708977] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return self.greenlet.switch() [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] result = function(*args, **kwargs) [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] return func(*args, **kwargs) [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] raise e [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] nwinfo = self.network_api.allocate_for_instance( [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] created_port_ids = self._update_ports_for_instance( [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] with excutils.save_and_reraise_exception(): [ 603.709353] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] self.force_reraise() [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] raise self.value [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] updated_port = self._update_port( [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] _ensure_no_port_binding_failure(port) [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] raise exception.PortBindingFailed(port_id=port['id']) [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] nova.exception.PortBindingFailed: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. [ 603.709680] env[61995]: ERROR nova.compute.manager [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] [ 603.710910] env[61995]: DEBUG nova.compute.utils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 603.714104] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Build of instance 15c8668b-655a-4e75-ae6b-4b892a04487a was re-scheduled: Binding failed for port 33fa9a01-a845-4b72-a99d-8cd1a7c7a063, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 603.714564] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 603.714800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Acquiring lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.714977] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Acquired lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.715107] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.717444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.097s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.717444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.717444] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 603.717822] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.822s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.720591] env[61995]: DEBUG nova.network.neutron [-] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.725019] env[61995]: DEBUG nova.network.neutron [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.726314] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588e9566-90bb-4862-9f44-5466c94a6f68 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.738202] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2662a71-08a8-4d2e-bbb5-4114312b4373 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.757781] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e95581-45d7-4b72-ae7f-a88daba65942 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.765778] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad341c1d-a125-46c1-847b-835279204baa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.804604] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181666MB free_disk=67GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 603.806597] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.050840] env[61995]: ERROR nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. [ 604.050840] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.050840] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.050840] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.050840] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.050840] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.050840] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.050840] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.050840] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.050840] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 604.050840] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.050840] env[61995]: ERROR nova.compute.manager raise self.value [ 604.050840] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.050840] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.050840] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.050840] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.051341] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.051341] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.051341] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. [ 604.051341] env[61995]: ERROR nova.compute.manager [ 604.051341] env[61995]: Traceback (most recent call last): [ 604.051341] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.051341] env[61995]: listener.cb(fileno) [ 604.051341] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.051341] env[61995]: result = function(*args, **kwargs) [ 604.051341] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.051341] env[61995]: return func(*args, **kwargs) [ 604.051341] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.051341] env[61995]: raise e [ 604.051341] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.051341] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 604.051341] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.051341] env[61995]: created_port_ids = self._update_ports_for_instance( [ 604.051341] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.051341] env[61995]: with excutils.save_and_reraise_exception(): [ 604.051341] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.051341] env[61995]: self.force_reraise() [ 604.051341] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.051341] env[61995]: raise self.value [ 604.051341] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.051341] env[61995]: updated_port = self._update_port( [ 604.051341] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.051341] env[61995]: _ensure_no_port_binding_failure(port) [ 604.051341] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.051341] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.052036] env[61995]: nova.exception.PortBindingFailed: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. [ 604.052036] env[61995]: Removing descriptor: 17 [ 604.052036] env[61995]: ERROR nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] Traceback (most recent call last): [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] yield resources [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self.driver.spawn(context, instance, image_meta, [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.052036] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] vm_ref = self.build_virtual_machine(instance, [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] for vif in network_info: [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return self._sync_wrapper(fn, *args, **kwargs) [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self.wait() [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self[:] = self._gt.wait() [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return self._exit_event.wait() [ 604.052398] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] result = hub.switch() [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return self.greenlet.switch() [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] result = function(*args, **kwargs) [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return func(*args, **kwargs) [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] raise e [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] nwinfo = self.network_api.allocate_for_instance( [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.052708] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] created_port_ids = self._update_ports_for_instance( [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] with excutils.save_and_reraise_exception(): [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self.force_reraise() [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] raise self.value [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] updated_port = self._update_port( [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] _ensure_no_port_binding_failure(port) [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.053050] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] raise exception.PortBindingFailed(port_id=port['id']) [ 604.053338] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] nova.exception.PortBindingFailed: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. [ 604.053338] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] [ 604.053338] env[61995]: INFO nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Terminating instance [ 604.054825] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.055015] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquired lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.055219] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.123216] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Releasing lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.123621] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.124209] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 604.124531] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1fb93ba6-3fe0-46c9-b17c-b90ca231e43b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.139984] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa0279f-635b-483a-a8eb-f53bbf9f35c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.162393] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 853dd1f7-e8f7-4d55-b53b-933ed8861b3d could not be found. [ 604.162731] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 604.162851] env[61995]: INFO nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 604.163247] env[61995]: DEBUG oslo.service.loopingcall [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.163362] env[61995]: DEBUG nova.compute.manager [-] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.163458] env[61995]: DEBUG nova.network.neutron [-] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.188328] env[61995]: DEBUG nova.network.neutron [-] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.265556] env[61995]: INFO nova.compute.manager [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: b6cb4bf3-291a-4aea-b64e-86597ad34dab] Took 1.09 seconds to deallocate network for instance. [ 604.268367] env[61995]: INFO nova.compute.manager [-] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Took 1.10 seconds to deallocate network for instance. [ 604.269556] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.280657] env[61995]: DEBUG nova.compute.claims [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.280657] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.376983] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Acquiring lock "3252888a-31ec-4d93-a8e2-b67aa0bab1e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.377689] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Lock "3252888a-31ec-4d93-a8e2-b67aa0bab1e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.412265] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.556917] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ee6e65-e235-414f-b158-153ac4084f8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.571703] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f114c7-3b7e-41ea-b2e0-1e334373e9a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.609399] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ff8056-3dd0-4570-8c3a-8cb65858003b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.618404] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4249754c-fdbf-462b-bdb1-7b24116f2c3c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.632114] env[61995]: DEBUG nova.compute.provider_tree [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.691189] env[61995]: DEBUG nova.network.neutron [-] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.758908] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.918768] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Releasing lock "refresh_cache-15c8668b-655a-4e75-ae6b-4b892a04487a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.918768] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 604.918768] env[61995]: DEBUG nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.918768] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.990036] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.138017] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.138017] env[61995]: DEBUG nova.scheduler.client.report [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.195770] env[61995]: INFO nova.compute.manager [-] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Took 1.03 seconds to deallocate network for instance. [ 605.205839] env[61995]: DEBUG nova.compute.claims [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.205839] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.218154] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "862694a5-63f8-4e2d-982a-329f4a071d29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.218409] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "862694a5-63f8-4e2d-982a-329f4a071d29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.317953] env[61995]: INFO nova.scheduler.client.report [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Deleted allocations for instance b6cb4bf3-291a-4aea-b64e-86597ad34dab [ 605.493893] env[61995]: DEBUG nova.network.neutron [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.648044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Releasing lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.648044] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.648044] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 605.648044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.648798] env[61995]: ERROR nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Traceback (most recent call last): [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self.driver.spawn(context, instance, image_meta, [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] vm_ref = self.build_virtual_machine(instance, [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.648798] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] for vif in network_info: [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return self._sync_wrapper(fn, *args, **kwargs) [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self.wait() [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self[:] = self._gt.wait() [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return self._exit_event.wait() [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] result = hub.switch() [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.649096] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return self.greenlet.switch() [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] result = function(*args, **kwargs) [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] return func(*args, **kwargs) [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] raise e [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] nwinfo = self.network_api.allocate_for_instance( [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] created_port_ids = self._update_ports_for_instance( [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] with excutils.save_and_reraise_exception(): [ 605.649415] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] self.force_reraise() [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] raise self.value [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] updated_port = self._update_port( [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] _ensure_no_port_binding_failure(port) [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] raise exception.PortBindingFailed(port_id=port['id']) [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] nova.exception.PortBindingFailed: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. [ 605.649725] env[61995]: ERROR nova.compute.manager [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] [ 605.649994] env[61995]: DEBUG nova.compute.utils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.651629] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf563d42-2452-4f45-9a71-468b3b5fabff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.653610] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Build of instance a7e67e95-3ac0-485a-94b9-59b68215ed7b was re-scheduled: Binding failed for port 521faecc-3dc6-4c06-8df0-f05b64e0d910, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.653967] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 605.654927] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Acquiring lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.654973] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Acquired lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.655199] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.658405] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.687s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.661557] env[61995]: INFO nova.compute.claims [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.674192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13908b39-f2e7-4be4-8635-54aa45d4d3ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.701570] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f59d6eee-6c40-4a99-b141-626e68308270 could not be found. [ 605.701698] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 605.701955] env[61995]: INFO nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Took 0.06 seconds to destroy the instance on the hypervisor. [ 605.702233] env[61995]: DEBUG oslo.service.loopingcall [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.702512] env[61995]: DEBUG nova.compute.manager [-] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.702618] env[61995]: DEBUG nova.network.neutron [-] [instance: f59d6eee-6c40-4a99-b141-626e68308270] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.733585] env[61995]: DEBUG nova.network.neutron [-] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.830244] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f5503ffd-f494-4a2f-82b7-1868435b943c tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "b6cb4bf3-291a-4aea-b64e-86597ad34dab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.502s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.849494] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Received event network-changed-ce3de92c-90ef-4b18-894b-8b9361218867 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 605.849810] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Refreshing instance network info cache due to event network-changed-ce3de92c-90ef-4b18-894b-8b9361218867. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 605.849995] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquiring lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.851514] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquired lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.851514] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Refreshing network info cache for port ce3de92c-90ef-4b18-894b-8b9361218867 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 605.997485] env[61995]: INFO nova.compute.manager [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] [instance: 15c8668b-655a-4e75-ae6b-4b892a04487a] Took 1.08 seconds to deallocate network for instance. [ 606.203127] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.239416] env[61995]: DEBUG nova.network.neutron [-] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.259427] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Acquiring lock "7bdf0b12-25d2-45a5-af57-b1bb31148f3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.260432] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Lock "7bdf0b12-25d2-45a5-af57-b1bb31148f3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.332602] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.359924] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.391195] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.482094] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.745271] env[61995]: INFO nova.compute.manager [-] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Took 1.04 seconds to deallocate network for instance. [ 606.748563] env[61995]: DEBUG nova.compute.claims [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.748761] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.859322] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.864516] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Releasing lock "refresh_cache-a7e67e95-3ac0-485a-94b9-59b68215ed7b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.864516] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.864845] env[61995]: DEBUG nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.864910] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.885861] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.985503] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Releasing lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.985770] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Received event network-vif-deleted-ce3de92c-90ef-4b18-894b-8b9361218867 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 606.985968] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Received event network-changed-2cfd262e-b05a-4c85-84ee-0c58efb14505 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 606.986157] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Refreshing instance network info cache due to event network-changed-2cfd262e-b05a-4c85-84ee-0c58efb14505. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 606.986364] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquiring lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.986504] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquired lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.986953] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Refreshing network info cache for port 2cfd262e-b05a-4c85-84ee-0c58efb14505 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.989121] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2321723-7f70-42df-ae51-8fa475752b47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.999435] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89240449-0a8e-4ac8-8e93-3bf4ad87ba1d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.032377] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc07105-a09b-436e-bc2c-909a0a2af6d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.041585] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27592bba-038c-4ab9-ac98-fab307eaa93f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.057648] env[61995]: DEBUG nova.compute.provider_tree [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.079786] env[61995]: INFO nova.scheduler.client.report [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Deleted allocations for instance 15c8668b-655a-4e75-ae6b-4b892a04487a [ 607.389716] env[61995]: DEBUG nova.network.neutron [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.521550] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.560416] env[61995]: DEBUG nova.scheduler.client.report [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.588269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-42c5e98d-9b4f-405e-bb62-b2cda66d48d6 tempest-ServerDiagnosticsNegativeTest-708859919 tempest-ServerDiagnosticsNegativeTest-708859919-project-member] Lock "15c8668b-655a-4e75-ae6b-4b892a04487a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.391s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.613098] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.892675] env[61995]: INFO nova.compute.manager [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] [instance: a7e67e95-3ac0-485a-94b9-59b68215ed7b] Took 1.03 seconds to deallocate network for instance. [ 608.066915] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.067555] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.071021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.709s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.090596] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.117080] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Releasing lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.117380] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Received event network-vif-deleted-2cfd262e-b05a-4c85-84ee-0c58efb14505 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.117598] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Received event network-changed-6e29aea8-9a11-47d4-9ac9-42a9899f4c13 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.118087] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Refreshing instance network info cache due to event network-changed-6e29aea8-9a11-47d4-9ac9-42a9899f4c13. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 608.118259] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquiring lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.118481] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquired lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.118666] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Refreshing network info cache for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 608.572597] env[61995]: DEBUG nova.compute.utils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.574417] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.574621] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 608.612679] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.711348] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.797043] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Acquiring lock "5ca64ce9-9856-4e65-a2e7-af11ae0211da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.797043] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Lock "5ca64ce9-9856-4e65-a2e7-af11ae0211da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.797330] env[61995]: DEBUG nova.policy [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a611017feca4f36a48d8fc2fd28d21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0bc85648bdc544ae868b7cce0ff66dd7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 608.863581] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faad07f0-2b76-4367-99a5-07c5f539bcb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.871492] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89b0652-ef4d-401d-9f3a-fbbf080b21ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.912768] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb478ebd-bbbe-49b5-b0cd-5e5c0401c30e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.917852] env[61995]: DEBUG nova.compute.manager [req-c78871a8-df48-430a-8e42-39fab67fdfc8 req-83c58ed4-e163-4fe9-ba96-8aebeaf9050b service nova] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Received event network-vif-deleted-a9b40936-87b7-45d7-a898-ccab46e8458c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.922693] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04481dd0-9856-4573-8c2f-8aac8229fe0f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.927486] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.941463] env[61995]: DEBUG nova.compute.provider_tree [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.946227] env[61995]: INFO nova.scheduler.client.report [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Deleted allocations for instance a7e67e95-3ac0-485a-94b9-59b68215ed7b [ 609.081965] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.379699] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Successfully created port: dc34f44e-6e15-4cc3-8f90-df87496e6907 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.434156] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Releasing lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.434156] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Received event network-vif-deleted-6e29aea8-9a11-47d4-9ac9-42a9899f4c13 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 609.434156] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Received event network-changed-a9b40936-87b7-45d7-a898-ccab46e8458c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 609.434156] env[61995]: DEBUG nova.compute.manager [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Refreshing instance network info cache due to event network-changed-a9b40936-87b7-45d7-a898-ccab46e8458c. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 609.434156] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquiring lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.434560] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Acquired lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.434560] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Refreshing network info cache for port a9b40936-87b7-45d7-a898-ccab46e8458c {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 609.444198] env[61995]: DEBUG nova.scheduler.client.report [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.455680] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f4fdc9dc-1366-4eda-8247-267de9725457 tempest-InstanceActionsTestJSON-1482085479 tempest-InstanceActionsTestJSON-1482085479-project-member] Lock "a7e67e95-3ac0-485a-94b9-59b68215ed7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.080s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.566272] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Acquiring lock "303d1ec5-cc8a-4f25-a36f-aec39e856766" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.566978] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Lock "303d1ec5-cc8a-4f25-a36f-aec39e856766" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.952991] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.953270] env[61995]: ERROR nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Traceback (most recent call last): [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self.driver.spawn(context, instance, image_meta, [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] vm_ref = self.build_virtual_machine(instance, [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.953270] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] for vif in network_info: [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return self._sync_wrapper(fn, *args, **kwargs) [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self.wait() [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self[:] = self._gt.wait() [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return self._exit_event.wait() [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] result = hub.switch() [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.953568] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return self.greenlet.switch() [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] result = function(*args, **kwargs) [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] return func(*args, **kwargs) [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] raise e [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] nwinfo = self.network_api.allocate_for_instance( [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] created_port_ids = self._update_ports_for_instance( [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] with excutils.save_and_reraise_exception(): [ 609.953907] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] self.force_reraise() [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] raise self.value [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] updated_port = self._update_port( [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] _ensure_no_port_binding_failure(port) [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] raise exception.PortBindingFailed(port_id=port['id']) [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] nova.exception.PortBindingFailed: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. [ 609.954243] env[61995]: ERROR nova.compute.manager [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] [ 609.954520] env[61995]: DEBUG nova.compute.utils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.957477] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.595s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.961035] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Build of instance 1e690534-1a87-47e6-b433-70e0598e79f2 was re-scheduled: Binding failed for port b93eab32-89d8-411f-8ffb-a08e16f210df, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.961144] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.961332] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.961488] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.961689] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.966551] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.968594] env[61995]: DEBUG nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.092907] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.110893] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "9f5024ce-5a48-4d27-a412-35149ee843f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.111144] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "9f5024ce-5a48-4d27-a412-35149ee843f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.130765] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.131211] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.131523] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.131645] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.131847] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.133094] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.133094] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.133094] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.133094] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.133094] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.133415] env[61995]: DEBUG nova.virt.hardware [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.134106] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f000bcb6-103a-42f4-9dd3-b2051f1b7801 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.143572] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a69695d-fbf5-4841-b0a6-44006f40ee8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.149804] env[61995]: DEBUG nova.network.neutron [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.505118] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.511345] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.671188] env[61995]: DEBUG oslo_concurrency.lockutils [req-019343ff-a5b6-4340-bbce-5a37aa0489b8 req-4583c69b-2f0a-4bf4-a43a-864fd989aeb3 service nova] Releasing lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.712111] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.856947] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89f3d1a-becd-4945-87ef-d9ba7bc0ed01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.875360] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dd742a-7359-41f0-819c-2243d70f3fd3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.920233] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf8a260-94b9-47f4-b9b1-d23adf9124ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.925153] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "06496d6a-324d-41d4-b3e5-52e3d93636bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.925389] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.933508] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd56334-8b45-4dec-9f1b-65290c886c2c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.949046] env[61995]: DEBUG nova.compute.provider_tree [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.217752] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "refresh_cache-1e690534-1a87-47e6-b433-70e0598e79f2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.218195] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.218471] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.218651] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.241319] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.278239] env[61995]: ERROR nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. [ 611.278239] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.278239] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.278239] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.278239] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.278239] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.278239] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.278239] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.278239] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.278239] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 611.278239] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.278239] env[61995]: ERROR nova.compute.manager raise self.value [ 611.278239] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.278239] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.278239] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.278239] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.278655] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.278655] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.278655] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. [ 611.278655] env[61995]: ERROR nova.compute.manager [ 611.278655] env[61995]: Traceback (most recent call last): [ 611.278655] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.278655] env[61995]: listener.cb(fileno) [ 611.278655] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.278655] env[61995]: result = function(*args, **kwargs) [ 611.278655] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.278655] env[61995]: return func(*args, **kwargs) [ 611.278655] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.278655] env[61995]: raise e [ 611.278655] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.278655] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 611.278655] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.278655] env[61995]: created_port_ids = self._update_ports_for_instance( [ 611.278655] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.278655] env[61995]: with excutils.save_and_reraise_exception(): [ 611.278655] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.278655] env[61995]: self.force_reraise() [ 611.278655] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.278655] env[61995]: raise self.value [ 611.278655] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.278655] env[61995]: updated_port = self._update_port( [ 611.278655] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.278655] env[61995]: _ensure_no_port_binding_failure(port) [ 611.278655] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.278655] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.279472] env[61995]: nova.exception.PortBindingFailed: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. [ 611.279472] env[61995]: Removing descriptor: 14 [ 611.279472] env[61995]: ERROR nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Traceback (most recent call last): [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] yield resources [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self.driver.spawn(context, instance, image_meta, [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.279472] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] vm_ref = self.build_virtual_machine(instance, [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] for vif in network_info: [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return self._sync_wrapper(fn, *args, **kwargs) [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self.wait() [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self[:] = self._gt.wait() [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return self._exit_event.wait() [ 611.279777] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] result = hub.switch() [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return self.greenlet.switch() [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] result = function(*args, **kwargs) [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return func(*args, **kwargs) [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] raise e [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] nwinfo = self.network_api.allocate_for_instance( [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.280138] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] created_port_ids = self._update_ports_for_instance( [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] with excutils.save_and_reraise_exception(): [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self.force_reraise() [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] raise self.value [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] updated_port = self._update_port( [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] _ensure_no_port_binding_failure(port) [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.280452] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] raise exception.PortBindingFailed(port_id=port['id']) [ 611.280745] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] nova.exception.PortBindingFailed: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. [ 611.280745] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] [ 611.280745] env[61995]: INFO nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Terminating instance [ 611.284676] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Acquiring lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.284676] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Acquired lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.284676] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.453579] env[61995]: DEBUG nova.scheduler.client.report [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.457858] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "d5a54c46-1a45-4bd5-8694-0defc757c455" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.457858] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.744741] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.815205] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.852028] env[61995]: DEBUG nova.compute.manager [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Received event network-changed-dc34f44e-6e15-4cc3-8f90-df87496e6907 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.852028] env[61995]: DEBUG nova.compute.manager [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Refreshing instance network info cache due to event network-changed-dc34f44e-6e15-4cc3-8f90-df87496e6907. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 611.852244] env[61995]: DEBUG oslo_concurrency.lockutils [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] Acquiring lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.958391] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.958791] env[61995]: ERROR nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Traceback (most recent call last): [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self.driver.spawn(context, instance, image_meta, [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] vm_ref = self.build_virtual_machine(instance, [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.958791] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] for vif in network_info: [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return self._sync_wrapper(fn, *args, **kwargs) [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self.wait() [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self[:] = self._gt.wait() [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return self._exit_event.wait() [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] result = hub.switch() [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.960050] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return self.greenlet.switch() [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] result = function(*args, **kwargs) [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] return func(*args, **kwargs) [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] raise e [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] nwinfo = self.network_api.allocate_for_instance( [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] created_port_ids = self._update_ports_for_instance( [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] with excutils.save_and_reraise_exception(): [ 611.960377] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] self.force_reraise() [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] raise self.value [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] updated_port = self._update_port( [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] _ensure_no_port_binding_failure(port) [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] raise exception.PortBindingFailed(port_id=port['id']) [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] nova.exception.PortBindingFailed: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. [ 611.960741] env[61995]: ERROR nova.compute.manager [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] [ 611.961018] env[61995]: DEBUG nova.compute.utils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.961018] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.036s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.966844] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Build of instance becd14dd-9ad6-4c5d-86b9-babd03dc0d92 was re-scheduled: Binding failed for port a8298ea3-78b5-4a6c-b651-4621f04fe498, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.967394] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.967394] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.967467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.967663] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 612.049037] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.249681] env[61995]: INFO nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 1e690534-1a87-47e6-b433-70e0598e79f2] Took 1.03 seconds to deallocate network for instance. [ 612.507800] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.556223] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Releasing lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.556629] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.556822] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 612.557654] env[61995]: DEBUG oslo_concurrency.lockutils [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] Acquired lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.557824] env[61995]: DEBUG nova.network.neutron [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Refreshing network info cache for port dc34f44e-6e15-4cc3-8f90-df87496e6907 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 612.558867] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89a4a331-d91d-4f8a-99b0-af81119c99ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.570546] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba4a36f-ec39-4881-994b-8eb6a8d3d9e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.604770] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bdc007d4-59d9-4fc8-8a63-c2ae0d058d48 could not be found. [ 612.605031] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 612.605556] env[61995]: INFO nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Took 0.05 seconds to destroy the instance on the hypervisor. [ 612.607540] env[61995]: DEBUG oslo.service.loopingcall [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.607540] env[61995]: DEBUG nova.compute.manager [-] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.607540] env[61995]: DEBUG nova.network.neutron [-] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 612.635652] env[61995]: DEBUG nova.network.neutron [-] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.678691] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.854289] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4d09c0-08b9-43c5-97d3-2fdf450222db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.863679] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e0743c-88e8-4c0c-8e9f-906a7bf5785d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.902974] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8562d80-6bf2-4f4e-831b-c33ed9f83b95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.911480] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880c76d4-8f7a-40aa-aac9-e8e3a287eafb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.933857] env[61995]: DEBUG nova.compute.provider_tree [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.096036] env[61995]: DEBUG nova.network.neutron [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.139875] env[61995]: DEBUG nova.network.neutron [-] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.181138] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "refresh_cache-becd14dd-9ad6-4c5d-86b9-babd03dc0d92" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.181418] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.181557] env[61995]: DEBUG nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.181726] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 613.283849] env[61995]: INFO nova.scheduler.client.report [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleted allocations for instance 1e690534-1a87-47e6-b433-70e0598e79f2 [ 613.384648] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.439100] env[61995]: DEBUG nova.scheduler.client.report [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.520446] env[61995]: DEBUG nova.network.neutron [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.642016] env[61995]: INFO nova.compute.manager [-] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Took 1.04 seconds to deallocate network for instance. [ 613.644760] env[61995]: DEBUG nova.compute.claims [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.644951] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.792147] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "1e690534-1a87-47e6-b433-70e0598e79f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.307s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.887801] env[61995]: DEBUG nova.network.neutron [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.942152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.942152] env[61995]: ERROR nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. [ 613.942152] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Traceback (most recent call last): [ 613.942152] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.942152] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self.driver.spawn(context, instance, image_meta, [ 613.942152] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.942152] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.942152] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.942152] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] vm_ref = self.build_virtual_machine(instance, [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] for vif in network_info: [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return self._sync_wrapper(fn, *args, **kwargs) [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self.wait() [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self[:] = self._gt.wait() [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return self._exit_event.wait() [ 613.943159] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] result = hub.switch() [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return self.greenlet.switch() [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] result = function(*args, **kwargs) [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] return func(*args, **kwargs) [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] raise e [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] nwinfo = self.network_api.allocate_for_instance( [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.943664] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] created_port_ids = self._update_ports_for_instance( [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] with excutils.save_and_reraise_exception(): [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] self.force_reraise() [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] raise self.value [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] updated_port = self._update_port( [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] _ensure_no_port_binding_failure(port) [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.944164] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] raise exception.PortBindingFailed(port_id=port['id']) [ 613.944621] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] nova.exception.PortBindingFailed: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. [ 613.944621] env[61995]: ERROR nova.compute.manager [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] [ 613.944621] env[61995]: DEBUG nova.compute.utils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.945907] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.140s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.947742] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Build of instance 9073413b-fc16-440b-84ad-6d8b826dc68a was re-scheduled: Binding failed for port ce3de92c-90ef-4b18-894b-8b9361218867, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.948884] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.949088] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Acquiring lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.949184] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Acquired lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.949381] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.023746] env[61995]: DEBUG oslo_concurrency.lockutils [req-e6e7b310-a7b7-4fd0-9c11-d7058cdd1eee req-d858ed4a-51c0-4169-a0a1-bfecaa08904f service nova] Releasing lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.251359] env[61995]: DEBUG nova.compute.manager [req-ecd2f447-9300-4759-9ea2-e01cf90d3369 req-5fc225dd-eef7-4c24-bac0-0619887e2029 service nova] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Received event network-vif-deleted-dc34f44e-6e15-4cc3-8f90-df87496e6907 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.294698] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.390600] env[61995]: INFO nova.compute.manager [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: becd14dd-9ad6-4c5d-86b9-babd03dc0d92] Took 1.21 seconds to deallocate network for instance. [ 614.496188] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.723467] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.841157] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.226676] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Releasing lock "refresh_cache-9073413b-fc16-440b-84ad-6d8b826dc68a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.226931] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.227117] env[61995]: DEBUG nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.228239] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.281569] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.450862] env[61995]: INFO nova.scheduler.client.report [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleted allocations for instance becd14dd-9ad6-4c5d-86b9-babd03dc0d92 [ 615.498947] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance becd14dd-9ad6-4c5d-86b9-babd03dc0d92 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 615.790099] env[61995]: DEBUG nova.network.neutron [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.961799] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e398b880-5c49-4914-ab37-3cdba3f0a31f tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "becd14dd-9ad6-4c5d-86b9-babd03dc0d92" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.427s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.003641] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 9073413b-fc16-440b-84ad-6d8b826dc68a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 616.003809] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ebd55d0b-c807-4adc-b1e4-b738f61b9bb3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 616.003941] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 853dd1f7-e8f7-4d55-b53b-933ed8861b3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 616.004079] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance f59d6eee-6c40-4a99-b141-626e68308270 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 616.004337] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance bdc007d4-59d9-4fc8-8a63-c2ae0d058d48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 616.292607] env[61995]: INFO nova.compute.manager [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] [instance: 9073413b-fc16-440b-84ad-6d8b826dc68a] Took 1.07 seconds to deallocate network for instance. [ 616.468131] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.511269] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ac219077-ed13-45ee-8d47-ff5dd5abe89a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 617.001142] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.013471] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 278cba4c-5bfb-42e0-b502-8f524ba1f380 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 617.338042] env[61995]: INFO nova.scheduler.client.report [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Deleted allocations for instance 9073413b-fc16-440b-84ad-6d8b826dc68a [ 617.520719] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 617.846467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-560f1584-68a9-41bf-863b-e0e171131310 tempest-ServerDiagnosticsTest-1387049184 tempest-ServerDiagnosticsTest-1387049184-project-member] Lock "9073413b-fc16-440b-84ad-6d8b826dc68a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.966s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.973034] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.973034] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.021640] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance adb60754-9217-4b1b-b5cd-98553ff5bbbf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 618.353260] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.528264] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 596463d7-04ca-4f29-8adf-c451310aaf15 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 618.896172] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.035021] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 3252888a-31ec-4d93-a8e2-b67aa0bab1e8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 619.127567] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.127723] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.176922] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.176922] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.536079] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 862694a5-63f8-4e2d-982a-329f4a071d29 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 620.042971] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 7bdf0b12-25d2-45a5-af57-b1bb31148f3d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 620.544761] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 5ca64ce9-9856-4e65-a2e7-af11ae0211da has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 620.982346] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "e53d10a3-1d16-498d-b4d8-abe7205410ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.982346] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.049771] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 303d1ec5-cc8a-4f25-a36f-aec39e856766 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 621.554970] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 9f5024ce-5a48-4d27-a412-35149ee843f3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 622.058881] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 06496d6a-324d-41d4-b3e5-52e3d93636bb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 622.087989] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.089228] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.564627] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d5a54c46-1a45-4bd5-8694-0defc757c455 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 622.564956] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 622.565021] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 622.888078] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1299882-90a9-42b1-a952-96af5ae5b468 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.897026] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e3b362-5591-4ca0-8e2a-53bf9bbb50e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.935838] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758342a6-f5db-416e-986a-af16cf0d62dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.944714] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0160b009-a4ab-4b4e-8512-1969a583d8e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.959197] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.463572] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.621032] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "1c4eb398-c4e9-4588-ba48-2805e3fee750" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.621748] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.792303] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.792547] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.970667] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 623.970951] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.027s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.971312] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.691s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.888588] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de42f247-1983-43e6-95f1-09bcfdf5fbe7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.897012] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e744ca-b7e7-448f-9b94-ca05178bd3ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.929088] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97cac32-9ca2-4cb1-8fce-fb29dc4690af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.936678] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3b07b5-ab42-466c-8186-1414cf5e1599 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.953328] env[61995]: DEBUG nova.compute.provider_tree [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.459762] env[61995]: DEBUG nova.scheduler.client.report [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.966251] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.966919] env[61995]: ERROR nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Traceback (most recent call last): [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self.driver.spawn(context, instance, image_meta, [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] vm_ref = self.build_virtual_machine(instance, [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.966919] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] for vif in network_info: [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return self._sync_wrapper(fn, *args, **kwargs) [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self.wait() [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self[:] = self._gt.wait() [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return self._exit_event.wait() [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] result = hub.switch() [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.967363] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return self.greenlet.switch() [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] result = function(*args, **kwargs) [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] return func(*args, **kwargs) [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] raise e [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] nwinfo = self.network_api.allocate_for_instance( [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] created_port_ids = self._update_ports_for_instance( [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] with excutils.save_and_reraise_exception(): [ 625.967723] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] self.force_reraise() [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] raise self.value [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] updated_port = self._update_port( [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] _ensure_no_port_binding_failure(port) [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] raise exception.PortBindingFailed(port_id=port['id']) [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] nova.exception.PortBindingFailed: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. [ 625.968166] env[61995]: ERROR nova.compute.manager [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] [ 625.968499] env[61995]: DEBUG nova.compute.utils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.970248] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Build of instance ebd55d0b-c807-4adc-b1e4-b738f61b9bb3 was re-scheduled: Binding failed for port 2cfd262e-b05a-4c85-84ee-0c58efb14505, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.970729] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.971045] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Acquiring lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.972268] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Acquired lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.972500] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.973919] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.768s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.511127] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.726921] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.925022] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267ad091-a042-4279-8f25-17cb5cb12322 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.933503] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c004dcb-e18c-4d2b-9525-0808ebcbe816 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.967808] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a302de64-6c6f-4220-9884-ab3be14fe34d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.975599] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3373f0f-7c6b-4de9-9be1-7466a4bd2477 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.993265] env[61995]: DEBUG nova.compute.provider_tree [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.230945] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Releasing lock "refresh_cache-ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.231304] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.231482] env[61995]: DEBUG nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.231650] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 627.280275] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.499303] env[61995]: DEBUG nova.scheduler.client.report [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.783459] env[61995]: DEBUG nova.network.neutron [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.003782] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.004707] env[61995]: ERROR nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Traceback (most recent call last): [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self.driver.spawn(context, instance, image_meta, [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] vm_ref = self.build_virtual_machine(instance, [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.004707] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] for vif in network_info: [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return self._sync_wrapper(fn, *args, **kwargs) [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self.wait() [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self[:] = self._gt.wait() [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return self._exit_event.wait() [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] result = hub.switch() [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.005155] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return self.greenlet.switch() [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] result = function(*args, **kwargs) [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] return func(*args, **kwargs) [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] raise e [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] nwinfo = self.network_api.allocate_for_instance( [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] created_port_ids = self._update_ports_for_instance( [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] with excutils.save_and_reraise_exception(): [ 628.005526] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] self.force_reraise() [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] raise self.value [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] updated_port = self._update_port( [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] _ensure_no_port_binding_failure(port) [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] raise exception.PortBindingFailed(port_id=port['id']) [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] nova.exception.PortBindingFailed: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. [ 628.005937] env[61995]: ERROR nova.compute.manager [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] [ 628.006297] env[61995]: DEBUG nova.compute.utils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.007011] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.258s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.011908] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Build of instance 853dd1f7-e8f7-4d55-b53b-933ed8861b3d was re-scheduled: Binding failed for port 6e29aea8-9a11-47d4-9ac9-42a9899f4c13, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.012420] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.012656] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Acquiring lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.012798] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Acquired lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.012956] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.287068] env[61995]: INFO nova.compute.manager [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] [instance: ebd55d0b-c807-4adc-b1e4-b738f61b9bb3] Took 1.06 seconds to deallocate network for instance. [ 628.562913] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.623105] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "45fdae21-bec2-438a-8f2a-c23b0e428057" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.623388] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.912777] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.942806] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e56b66-af26-469e-9fda-3492bf516030 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.955335] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bb19e1-a599-4709-a1c7-108961d8f0c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.004897] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2b06e9-d7e9-476d-a658-cf997410b281 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.012708] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eef8788-a528-459a-b338-4936b1c59c86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.030174] env[61995]: DEBUG nova.compute.provider_tree [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.334804] env[61995]: INFO nova.scheduler.client.report [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Deleted allocations for instance ebd55d0b-c807-4adc-b1e4-b738f61b9bb3 [ 629.414454] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Releasing lock "refresh_cache-853dd1f7-e8f7-4d55-b53b-933ed8861b3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.414743] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.414951] env[61995]: DEBUG nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.415184] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.450019] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.533458] env[61995]: DEBUG nova.scheduler.client.report [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.842914] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7e226a81-84e1-4b9a-a840-94a7edd73c41 tempest-ServersAdminNegativeTestJSON-1645515566 tempest-ServersAdminNegativeTestJSON-1645515566-project-member] Lock "ebd55d0b-c807-4adc-b1e4-b738f61b9bb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.953890] env[61995]: DEBUG nova.network.neutron [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.990069] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.990069] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.041262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.041262] env[61995]: ERROR nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. [ 630.041262] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] Traceback (most recent call last): [ 630.041262] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.041262] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self.driver.spawn(context, instance, image_meta, [ 630.041262] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 630.041262] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.041262] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.041262] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] vm_ref = self.build_virtual_machine(instance, [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] for vif in network_info: [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return self._sync_wrapper(fn, *args, **kwargs) [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self.wait() [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self[:] = self._gt.wait() [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return self._exit_event.wait() [ 630.041729] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] result = hub.switch() [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return self.greenlet.switch() [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] result = function(*args, **kwargs) [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] return func(*args, **kwargs) [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] raise e [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] nwinfo = self.network_api.allocate_for_instance( [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.042093] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] created_port_ids = self._update_ports_for_instance( [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] with excutils.save_and_reraise_exception(): [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] self.force_reraise() [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] raise self.value [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] updated_port = self._update_port( [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] _ensure_no_port_binding_failure(port) [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.042438] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] raise exception.PortBindingFailed(port_id=port['id']) [ 630.042766] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] nova.exception.PortBindingFailed: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. [ 630.042766] env[61995]: ERROR nova.compute.manager [instance: f59d6eee-6c40-4a99-b141-626e68308270] [ 630.042766] env[61995]: DEBUG nova.compute.utils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.042766] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.183s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.044653] env[61995]: INFO nova.compute.claims [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.048031] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Build of instance f59d6eee-6c40-4a99-b141-626e68308270 was re-scheduled: Binding failed for port a9b40936-87b7-45d7-a898-ccab46e8458c, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.049394] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.049670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.049816] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquired lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.050232] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.347919] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.459748] env[61995]: INFO nova.compute.manager [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] [instance: 853dd1f7-e8f7-4d55-b53b-933ed8861b3d] Took 1.04 seconds to deallocate network for instance. [ 630.590211] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.848187] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.884812] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.351324] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Releasing lock "refresh_cache-f59d6eee-6c40-4a99-b141-626e68308270" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.351587] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.351748] env[61995]: DEBUG nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.351946] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.378599] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.475827] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0f6cd2-fa32-449e-982d-5938da6829b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.487063] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc77ee39-4260-448c-a98d-128b568c3ca6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.524024] env[61995]: INFO nova.scheduler.client.report [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Deleted allocations for instance 853dd1f7-e8f7-4d55-b53b-933ed8861b3d [ 631.531726] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9914bda-87bb-4807-8db4-3d198e7f927e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.540650] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d368e3-1620-462a-bffd-47c5c27c5d86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.558338] env[61995]: DEBUG nova.compute.provider_tree [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.882684] env[61995]: DEBUG nova.network.neutron [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.008037] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.008037] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.037851] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7867c4a5-5cc0-4b43-88d7-1e6ad39b22ea tempest-ServerRescueTestJSON-1570964489 tempest-ServerRescueTestJSON-1570964489-project-member] Lock "853dd1f7-e8f7-4d55-b53b-933ed8861b3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.262s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.061887] env[61995]: DEBUG nova.scheduler.client.report [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.386697] env[61995]: INFO nova.compute.manager [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: f59d6eee-6c40-4a99-b141-626e68308270] Took 1.03 seconds to deallocate network for instance. [ 632.546782] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.567825] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.568666] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.572613] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.959s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.573016] env[61995]: INFO nova.compute.claims [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.695032] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "c350194c-ed3e-4996-871a-70b4cb99c7db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.695332] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.077261] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.078616] env[61995]: DEBUG nova.compute.utils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.083021] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.090816] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 633.149689] env[61995]: DEBUG nova.policy [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51e4b19f6bf5459eaa74defb07618c64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30faef7d43954cf98a83ad71a84c9ed5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 633.437620] env[61995]: INFO nova.scheduler.client.report [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Deleted allocations for instance f59d6eee-6c40-4a99-b141-626e68308270 [ 633.584740] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 633.683182] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Successfully created port: 5c8b7665-348d-4130-bf59-fbf14180b7e0 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.953678] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df2f6318-0a2a-4eb3-bdc7-7a8afa306ee0 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "f59d6eee-6c40-4a99-b141-626e68308270" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.824s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.028062] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931dfcc5-fc78-4fa4-9b86-efd50b1788ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.038185] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b985e9db-412b-495e-a5c5-002aedf12b80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.070851] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855a7852-d605-4792-953a-e667f5a756dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.078474] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6766df-abd7-4f40-9f0f-ee7578b74990 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.092830] env[61995]: DEBUG nova.compute.provider_tree [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.459567] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.602391] env[61995]: DEBUG nova.scheduler.client.report [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.606921] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 634.645250] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.645642] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.645642] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.645824] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.646228] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.646454] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.646748] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.646859] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.647037] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.647211] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.647896] env[61995]: DEBUG nova.virt.hardware [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.649842] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e61605b-4190-47f9-8117-855ed75c5e05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.661227] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc163ec3-586f-42c5-b4e2-9e232aaae487 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.993568] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.109330] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.114029] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.114993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.604s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.118420] env[61995]: INFO nova.compute.claims [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.624146] env[61995]: DEBUG nova.compute.utils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.627926] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 635.628466] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 635.730718] env[61995]: DEBUG nova.compute.manager [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Received event network-changed-5c8b7665-348d-4130-bf59-fbf14180b7e0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.730914] env[61995]: DEBUG nova.compute.manager [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Refreshing instance network info cache due to event network-changed-5c8b7665-348d-4130-bf59-fbf14180b7e0. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 635.731140] env[61995]: DEBUG oslo_concurrency.lockutils [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] Acquiring lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.731272] env[61995]: DEBUG oslo_concurrency.lockutils [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] Acquired lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.731424] env[61995]: DEBUG nova.network.neutron [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Refreshing network info cache for port 5c8b7665-348d-4130-bf59-fbf14180b7e0 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 635.739366] env[61995]: DEBUG nova.policy [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91fb46e100384b0fa5ec36465140c2ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f4dee4fa81a4fb9baa6bb00c305d3c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 635.996048] env[61995]: ERROR nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. [ 635.996048] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.996048] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.996048] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.996048] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.996048] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.996048] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.996048] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.996048] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.996048] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 635.996048] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.996048] env[61995]: ERROR nova.compute.manager raise self.value [ 635.996048] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.996048] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.996048] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.996048] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.996742] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.996742] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.996742] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. [ 635.996742] env[61995]: ERROR nova.compute.manager [ 635.996742] env[61995]: Traceback (most recent call last): [ 635.996742] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.996742] env[61995]: listener.cb(fileno) [ 635.996742] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.996742] env[61995]: result = function(*args, **kwargs) [ 635.996742] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.996742] env[61995]: return func(*args, **kwargs) [ 635.996742] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.996742] env[61995]: raise e [ 635.996742] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.996742] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 635.996742] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.996742] env[61995]: created_port_ids = self._update_ports_for_instance( [ 635.996742] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.996742] env[61995]: with excutils.save_and_reraise_exception(): [ 635.996742] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.996742] env[61995]: self.force_reraise() [ 635.996742] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.996742] env[61995]: raise self.value [ 635.996742] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.996742] env[61995]: updated_port = self._update_port( [ 635.996742] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.996742] env[61995]: _ensure_no_port_binding_failure(port) [ 635.996742] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.996742] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.997545] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. [ 635.997545] env[61995]: Removing descriptor: 14 [ 635.997545] env[61995]: ERROR nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Traceback (most recent call last): [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] yield resources [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self.driver.spawn(context, instance, image_meta, [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.997545] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] vm_ref = self.build_virtual_machine(instance, [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] for vif in network_info: [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return self._sync_wrapper(fn, *args, **kwargs) [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self.wait() [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self[:] = self._gt.wait() [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return self._exit_event.wait() [ 635.998448] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] result = hub.switch() [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return self.greenlet.switch() [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] result = function(*args, **kwargs) [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return func(*args, **kwargs) [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] raise e [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] nwinfo = self.network_api.allocate_for_instance( [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.998838] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] created_port_ids = self._update_ports_for_instance( [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] with excutils.save_and_reraise_exception(): [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self.force_reraise() [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] raise self.value [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] updated_port = self._update_port( [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] _ensure_no_port_binding_failure(port) [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.999269] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] raise exception.PortBindingFailed(port_id=port['id']) [ 635.999627] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] nova.exception.PortBindingFailed: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. [ 635.999627] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] [ 635.999627] env[61995]: INFO nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Terminating instance [ 636.001201] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Acquiring lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.128557] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.263433] env[61995]: DEBUG nova.network.neutron [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.430480] env[61995]: DEBUG nova.network.neutron [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.484390] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Successfully created port: 80ba18c9-8004-46b7-ae1a-fc17685d183e {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.534822] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a317057b-c172-4a2c-a1ca-17ea99e2e439 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.542806] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221f7d98-8e10-437a-9514-e0a4c200d54b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.584639] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5cb147-1ebd-4657-90b5-79898272a4e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.599633] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03600fe-a0f5-4c58-8272-96960ce47da7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.613574] env[61995]: DEBUG nova.compute.provider_tree [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.933553] env[61995]: DEBUG oslo_concurrency.lockutils [req-fe637290-e8d9-4b4d-b5e6-77b601a2d199 req-50ab6acb-db93-4d48-ac69-2d6c282fdbe1 service nova] Releasing lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.934085] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Acquired lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.934399] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.102786] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.103085] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.116243] env[61995]: DEBUG nova.scheduler.client.report [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.142284] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.171793] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.172051] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.172215] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.172399] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.172540] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.172683] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.172926] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.173190] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.173793] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.173793] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.173793] env[61995]: DEBUG nova.virt.hardware [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.174707] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9465d7-974c-4882-8b2f-a93031ca12de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.183529] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18d3a50-e74f-459b-a305-77a3ba1e6190 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.433284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.433908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.480131] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.622607] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.623461] env[61995]: DEBUG nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.631369] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.986s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.644012] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.127771] env[61995]: ERROR nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. [ 638.127771] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.127771] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.127771] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.127771] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.127771] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.127771] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.127771] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.127771] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.127771] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 638.127771] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.127771] env[61995]: ERROR nova.compute.manager raise self.value [ 638.127771] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.127771] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.127771] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.127771] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.128527] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.128527] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.128527] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. [ 638.128527] env[61995]: ERROR nova.compute.manager [ 638.128527] env[61995]: Traceback (most recent call last): [ 638.128527] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.128527] env[61995]: listener.cb(fileno) [ 638.128527] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.128527] env[61995]: result = function(*args, **kwargs) [ 638.128527] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.128527] env[61995]: return func(*args, **kwargs) [ 638.128527] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.128527] env[61995]: raise e [ 638.128527] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.128527] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 638.128527] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.128527] env[61995]: created_port_ids = self._update_ports_for_instance( [ 638.128527] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.128527] env[61995]: with excutils.save_and_reraise_exception(): [ 638.128527] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.128527] env[61995]: self.force_reraise() [ 638.128527] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.128527] env[61995]: raise self.value [ 638.128527] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.128527] env[61995]: updated_port = self._update_port( [ 638.128527] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.128527] env[61995]: _ensure_no_port_binding_failure(port) [ 638.128527] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.128527] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.132756] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. [ 638.132756] env[61995]: Removing descriptor: 17 [ 638.132756] env[61995]: ERROR nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Traceback (most recent call last): [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] yield resources [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self.driver.spawn(context, instance, image_meta, [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.132756] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] vm_ref = self.build_virtual_machine(instance, [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] for vif in network_info: [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return self._sync_wrapper(fn, *args, **kwargs) [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self.wait() [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self[:] = self._gt.wait() [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return self._exit_event.wait() [ 638.133385] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] result = hub.switch() [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return self.greenlet.switch() [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] result = function(*args, **kwargs) [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return func(*args, **kwargs) [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] raise e [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] nwinfo = self.network_api.allocate_for_instance( [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.133873] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] created_port_ids = self._update_ports_for_instance( [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] with excutils.save_and_reraise_exception(): [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self.force_reraise() [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] raise self.value [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] updated_port = self._update_port( [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] _ensure_no_port_binding_failure(port) [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.134433] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] raise exception.PortBindingFailed(port_id=port['id']) [ 638.134789] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] nova.exception.PortBindingFailed: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. [ 638.134789] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] [ 638.134789] env[61995]: INFO nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Terminating instance [ 638.134789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Acquiring lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.134789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Acquired lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.134789] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 638.140521] env[61995]: DEBUG nova.compute.utils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.140521] env[61995]: DEBUG nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Not allocating networking since 'none' was specified. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 638.147635] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Releasing lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.148087] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.148294] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 638.149030] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7eab41b3-6452-46b3-aab7-28a7aa3302cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.159111] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10cc63fe-b9db-485a-a15c-4548f3d5685d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.187724] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ac219077-ed13-45ee-8d47-ff5dd5abe89a could not be found. [ 638.187982] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 638.188233] env[61995]: INFO nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 638.188478] env[61995]: DEBUG oslo.service.loopingcall [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.191137] env[61995]: DEBUG nova.compute.manager [-] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.191242] env[61995]: DEBUG nova.network.neutron [-] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.236752] env[61995]: DEBUG nova.network.neutron [-] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.261374] env[61995]: DEBUG nova.compute.manager [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Received event network-vif-deleted-5c8b7665-348d-4130-bf59-fbf14180b7e0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.261471] env[61995]: DEBUG nova.compute.manager [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Received event network-changed-80ba18c9-8004-46b7-ae1a-fc17685d183e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.261667] env[61995]: DEBUG nova.compute.manager [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Refreshing instance network info cache due to event network-changed-80ba18c9-8004-46b7-ae1a-fc17685d183e. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 638.261750] env[61995]: DEBUG oslo_concurrency.lockutils [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] Acquiring lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.620680] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6eb8d15-87d5-41da-b852-3148348eb9c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.632083] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee2b280-cb17-4c77-aec6-0e83ca453006 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.665226] env[61995]: DEBUG nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.671021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777d6b95-bbee-4a0b-be25-a5bed9dc7264 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.675613] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.675837] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.681385] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8d6f94-8618-4b07-83e5-57baff38ffce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.685896] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.697981] env[61995]: DEBUG nova.compute.provider_tree [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.738135] env[61995]: DEBUG nova.network.neutron [-] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.777603] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.201250] env[61995]: DEBUG nova.scheduler.client.report [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.242177] env[61995]: INFO nova.compute.manager [-] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Took 1.05 seconds to deallocate network for instance. [ 639.243729] env[61995]: DEBUG nova.compute.claims [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.243729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.286018] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Releasing lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.286018] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.286018] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 639.286018] env[61995]: DEBUG oslo_concurrency.lockutils [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] Acquired lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.286018] env[61995]: DEBUG nova.network.neutron [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Refreshing network info cache for port 80ba18c9-8004-46b7-ae1a-fc17685d183e {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 639.286385] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64deb6e9-dfd0-42c1-952a-25142e258dac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.302281] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c613b4bd-369c-4e3b-a340-d2b0ec6e92fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.328458] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 278cba4c-5bfb-42e0-b502-8f524ba1f380 could not be found. [ 639.328740] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 639.328922] env[61995]: INFO nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Took 0.05 seconds to destroy the instance on the hypervisor. [ 639.329247] env[61995]: DEBUG oslo.service.loopingcall [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.329922] env[61995]: DEBUG nova.compute.manager [-] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.329922] env[61995]: DEBUG nova.network.neutron [-] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 639.353957] env[61995]: DEBUG nova.network.neutron [-] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.446109] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "55555deb-9883-483b-a901-42b1a66c60d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.446511] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "55555deb-9883-483b-a901-42b1a66c60d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.675074] env[61995]: DEBUG nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.705879] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.705879] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.706382] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.706382] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.708017] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.708017] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.708017] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.708017] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.708017] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.708301] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.708301] env[61995]: DEBUG nova.virt.hardware [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.708717] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.077s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.709337] env[61995]: ERROR nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Traceback (most recent call last): [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self.driver.spawn(context, instance, image_meta, [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] vm_ref = self.build_virtual_machine(instance, [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.709337] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] for vif in network_info: [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return self._sync_wrapper(fn, *args, **kwargs) [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self.wait() [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self[:] = self._gt.wait() [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return self._exit_event.wait() [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] result = hub.switch() [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.709728] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return self.greenlet.switch() [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] result = function(*args, **kwargs) [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] return func(*args, **kwargs) [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] raise e [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] nwinfo = self.network_api.allocate_for_instance( [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] created_port_ids = self._update_ports_for_instance( [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] with excutils.save_and_reraise_exception(): [ 639.710113] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] self.force_reraise() [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] raise self.value [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] updated_port = self._update_port( [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] _ensure_no_port_binding_failure(port) [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] raise exception.PortBindingFailed(port_id=port['id']) [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] nova.exception.PortBindingFailed: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. [ 639.710469] env[61995]: ERROR nova.compute.manager [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] [ 639.710862] env[61995]: DEBUG nova.compute.utils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.712241] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6282f3-45e0-419f-99a6-162ad947539a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.714601] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Build of instance bdc007d4-59d9-4fc8-8a63-c2ae0d058d48 was re-scheduled: Binding failed for port dc34f44e-6e15-4cc3-8f90-df87496e6907, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.715181] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 639.715347] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Acquiring lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.715405] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Acquired lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.715549] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.716671] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.876s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.718486] env[61995]: INFO nova.compute.claims [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.727987] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787ad8cf-4399-4167-a067-a3a572ac34a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.746998] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.756623] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.757307] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-387539fb-bf58-4f94-895f-ab26ddbf7cce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.770215] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Created folder: OpenStack in parent group-v4. [ 639.770398] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Creating folder: Project (c42710db14154be3bcfb761b7e309702). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.770655] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4d7673c-acd4-4a90-84f3-9daa0de8daff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.781153] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Created folder: Project (c42710db14154be3bcfb761b7e309702) in parent group-v185203. [ 639.781352] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Creating folder: Instances. Parent ref: group-v185204. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.781581] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83c1de13-2761-406d-8a00-e2530ad86ea1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.795604] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Created folder: Instances in parent group-v185204. [ 639.795751] env[61995]: DEBUG oslo.service.loopingcall [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.795989] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 639.796502] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-989815ba-5370-4ce1-88eb-94432fdaf5e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.815742] env[61995]: DEBUG nova.network.neutron [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.822648] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.822648] env[61995]: value = "task-794160" [ 639.822648] env[61995]: _type = "Task" [ 639.822648] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.833868] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794160, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.855541] env[61995]: DEBUG nova.network.neutron [-] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.894290] env[61995]: DEBUG nova.network.neutron [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.221885] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.222237] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.253148] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.337127] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794160, 'name': CreateVM_Task, 'duration_secs': 0.292561} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.340780] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 640.340780] env[61995]: DEBUG oslo_vmware.service [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764c237e-42dd-4d47-9213-b159c3a85dcf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.346943] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.347123] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.347854] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 640.348090] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4596b8f0-3daf-4711-ab6d-695fc8ee1255 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.352689] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 640.352689] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522393fd-3a04-e7f0-cf34-d13cece600c6" [ 640.352689] env[61995]: _type = "Task" [ 640.352689] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.359076] env[61995]: INFO nova.compute.manager [-] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Took 1.03 seconds to deallocate network for instance. [ 640.360039] env[61995]: DEBUG nova.compute.manager [req-ac2b0e8c-d915-44a3-9cb7-9afc3add2e3e req-420e258e-8731-47ae-8e18-a3362dea610f service nova] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Received event network-vif-deleted-80ba18c9-8004-46b7-ae1a-fc17685d183e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 640.362830] env[61995]: DEBUG nova.compute.claims [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.362830] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.366946] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.367175] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.367419] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.369382] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.369382] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.369382] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3275fd2-0157-485a-a6cc-38a10afa852c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.372928] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.376517] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.376699] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 640.377454] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0056050-74d1-4e37-a364-ea35a1d050c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.386620] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06fda715-ff1d-469f-8764-958552188f42 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.394121] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 640.394121] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5266d5d7-83d5-f7dc-d196-60a7d7c1dc84" [ 640.394121] env[61995]: _type = "Task" [ 640.394121] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.397747] env[61995]: DEBUG oslo_concurrency.lockutils [req-5420a8ad-2fb3-431c-942b-fde1901af17f req-6eceebad-369e-47a5-bc59-9f0706dcb4e9 service nova] Releasing lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.407400] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Preparing fetch location {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 640.407647] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Creating directory with path [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.407864] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1275b4d4-8ddb-4c60-9c7d-869edc8be58e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.427816] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Created directory with path [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.427956] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Fetch image to [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 640.428151] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Downloading image file data c1739132-aff3-4644-9522-fc02cf765bc8 to [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk on the data store datastore2 {{(pid=61995) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 640.428997] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f368b73d-c7a9-48bb-91db-f5362cbd2c00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.437048] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8af2df9-e339-4996-b760-8a36f55fc93f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.446432] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b37678-7cee-4f7d-83be-54d4e5e77477 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.477320] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618dda78-9777-4b4d-9a99-2df134828a6e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.483659] env[61995]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-a3652931-fe9a-4452-bf40-890ffc1c5d9f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.505598] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Downloading image file data c1739132-aff3-4644-9522-fc02cf765bc8 to the data store datastore2 {{(pid=61995) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 640.584196] env[61995]: DEBUG oslo_vmware.rw_handles [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61995) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 640.876687] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Releasing lock "refresh_cache-bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.876687] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.876687] env[61995]: DEBUG nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.877105] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.915460] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.286226] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de80e2b-dadd-4432-a000-ef071a28c482 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.298017] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e490c0a-20dd-482e-aba7-00fd4200d0d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.337306] env[61995]: DEBUG oslo_vmware.rw_handles [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Completed reading data from the image iterator. {{(pid=61995) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 641.337529] env[61995]: DEBUG oslo_vmware.rw_handles [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 641.338448] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1c6894-a5e9-4693-8ba3-e65b394c9748 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.346481] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9896be4-27e6-4bb3-a98d-e7997ddabd0d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.360627] env[61995]: DEBUG nova.compute.provider_tree [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.420743] env[61995]: DEBUG nova.network.neutron [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.468110] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Downloaded image file data c1739132-aff3-4644-9522-fc02cf765bc8 to vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk on the data store datastore2 {{(pid=61995) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 641.470163] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Caching image {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 641.470429] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Copying Virtual Disk [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk to [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 641.471010] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-41c362b9-ea73-45bc-8244-515b7eaf1e81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.478846] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 641.478846] env[61995]: value = "task-794161" [ 641.478846] env[61995]: _type = "Task" [ 641.478846] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.486874] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794161, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.863966] env[61995]: DEBUG nova.scheduler.client.report [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.923652] env[61995]: INFO nova.compute.manager [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] [instance: bdc007d4-59d9-4fc8-8a63-c2ae0d058d48] Took 1.05 seconds to deallocate network for instance. [ 641.990826] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794161, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.369221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.652s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.369544] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.372594] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.372s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.374278] env[61995]: INFO nova.compute.claims [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.489600] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794161, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675358} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.489810] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Copied Virtual Disk [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk to [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 642.489979] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleting the datastore file [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 642.490233] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ace4ad6f-74f0-4752-95e9-edc7208ed28a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.497589] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 642.497589] env[61995]: value = "task-794162" [ 642.497589] env[61995]: _type = "Task" [ 642.497589] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.506155] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.880376] env[61995]: DEBUG nova.compute.utils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 642.887716] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 642.887936] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 642.937081] env[61995]: DEBUG nova.policy [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d68919fd67f4c1993c5f108f4ea4b18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7272aa9c8ba04df892d48a4b18b575a9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 642.977021] env[61995]: INFO nova.scheduler.client.report [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Deleted allocations for instance bdc007d4-59d9-4fc8-8a63-c2ae0d058d48 [ 643.011443] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024313} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.012074] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 643.013621] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Moving file from [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85/c1739132-aff3-4644-9522-fc02cf765bc8 to [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8. {{(pid=61995) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 643.013621] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-06e55954-7594-42fe-a234-cc80006baf90 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.021547] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 643.021547] env[61995]: value = "task-794163" [ 643.021547] env[61995]: _type = "Task" [ 643.021547] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.030180] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794163, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.239336] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Successfully created port: 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.389761] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 643.483991] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a522acf7-7827-46ca-85a8-32315196fad4 tempest-VolumesAssistedSnapshotsTest-463282367 tempest-VolumesAssistedSnapshotsTest-463282367-project-member] Lock "bdc007d4-59d9-4fc8-8a63-c2ae0d058d48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.551s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.534506] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794163, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025265} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.534770] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] File moved {{(pid=61995) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 643.534963] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Cleaning up location [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 643.535384] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleting the datastore file [datastore2] vmware_temp/0d2e7c80-dc84-4f9c-8886-b88169089e85 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.535384] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c33fddec-7822-4474-bc16-f40448c57478 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.542068] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 643.542068] env[61995]: value = "task-794164" [ 643.542068] env[61995]: _type = "Task" [ 643.542068] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.553663] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.879250] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4786e93-c0d2-4bee-bf0d-ca31b9c753c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.888303] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57ce198-a647-4867-be27-fabbd817c252 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.935235] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e400154-222c-4713-b035-f6e9334d8821 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.942415] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cac6c45-f9a0-4c69-99c1-b7d5028c93ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.956638] env[61995]: DEBUG nova.compute.provider_tree [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.986077] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.057374] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034292} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.057768] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 644.058606] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a815adce-dd18-4f2d-924e-ef294bf1196c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.065966] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 644.065966] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527d2963-32d8-4760-7012-a27cbb677565" [ 644.065966] env[61995]: _type = "Task" [ 644.065966] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.073329] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527d2963-32d8-4760-7012-a27cbb677565, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.280770] env[61995]: DEBUG nova.compute.manager [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Received event network-changed-9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 644.280770] env[61995]: DEBUG nova.compute.manager [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Refreshing instance network info cache due to event network-changed-9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 644.280770] env[61995]: DEBUG oslo_concurrency.lockutils [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] Acquiring lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.280770] env[61995]: DEBUG oslo_concurrency.lockutils [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] Acquired lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.280770] env[61995]: DEBUG nova.network.neutron [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Refreshing network info cache for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 644.355822] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.356156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.442285] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 644.459593] env[61995]: DEBUG nova.scheduler.client.report [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.478314] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 644.479091] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 644.479091] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.479091] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 644.479091] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.479258] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 644.479387] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 644.479549] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 644.479764] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 644.479929] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 644.480120] env[61995]: DEBUG nova.virt.hardware [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.481087] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a40bb64-9745-49b4-9f09-741fe1fee2e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.490552] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0677f12-95ae-4ea9-a0f6-29e5b062927d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.495185] env[61995]: ERROR nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. [ 644.495185] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.495185] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.495185] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.495185] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.495185] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.495185] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.495185] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.495185] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.495185] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 644.495185] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.495185] env[61995]: ERROR nova.compute.manager raise self.value [ 644.495185] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.495185] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.495185] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.495185] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.495793] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.495793] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.495793] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. [ 644.495793] env[61995]: ERROR nova.compute.manager [ 644.495793] env[61995]: Traceback (most recent call last): [ 644.495793] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.495793] env[61995]: listener.cb(fileno) [ 644.495793] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.495793] env[61995]: result = function(*args, **kwargs) [ 644.495793] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.495793] env[61995]: return func(*args, **kwargs) [ 644.495793] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.495793] env[61995]: raise e [ 644.495793] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.495793] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 644.495793] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.495793] env[61995]: created_port_ids = self._update_ports_for_instance( [ 644.495793] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.495793] env[61995]: with excutils.save_and_reraise_exception(): [ 644.495793] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.495793] env[61995]: self.force_reraise() [ 644.495793] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.495793] env[61995]: raise self.value [ 644.495793] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.495793] env[61995]: updated_port = self._update_port( [ 644.495793] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.495793] env[61995]: _ensure_no_port_binding_failure(port) [ 644.495793] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.495793] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.497130] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. [ 644.497130] env[61995]: Removing descriptor: 17 [ 644.510279] env[61995]: ERROR nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Traceback (most recent call last): [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] yield resources [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self.driver.spawn(context, instance, image_meta, [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] vm_ref = self.build_virtual_machine(instance, [ 644.510279] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] for vif in network_info: [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] return self._sync_wrapper(fn, *args, **kwargs) [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self.wait() [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self[:] = self._gt.wait() [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] return self._exit_event.wait() [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 644.510775] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] current.throw(*self._exc) [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] result = function(*args, **kwargs) [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] return func(*args, **kwargs) [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] raise e [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] nwinfo = self.network_api.allocate_for_instance( [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] created_port_ids = self._update_ports_for_instance( [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] with excutils.save_and_reraise_exception(): [ 644.511160] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self.force_reraise() [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] raise self.value [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] updated_port = self._update_port( [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] _ensure_no_port_binding_failure(port) [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] raise exception.PortBindingFailed(port_id=port['id']) [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] nova.exception.PortBindingFailed: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. [ 644.511528] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] [ 644.511528] env[61995]: INFO nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Terminating instance [ 644.513497] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Acquiring lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.517767] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.575344] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527d2963-32d8-4760-7012-a27cbb677565, 'name': SearchDatastore_Task, 'duration_secs': 0.009274} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.575681] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.575957] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 644.576761] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0107d7e6-6776-4a9a-88cc-82d6a76c35c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.583176] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 644.583176] env[61995]: value = "task-794165" [ 644.583176] env[61995]: _type = "Task" [ 644.583176] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.591994] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.803450] env[61995]: DEBUG nova.network.neutron [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.919190] env[61995]: DEBUG nova.network.neutron [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.965472] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.966112] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.970112] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.076s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.971696] env[61995]: INFO nova.compute.claims [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.094633] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464585} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.094807] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 645.094989] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 645.095818] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca15204e-f201-42b9-941a-bc5216f665f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.104018] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 645.104018] env[61995]: value = "task-794166" [ 645.104018] env[61995]: _type = "Task" [ 645.104018] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.118971] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794166, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.423201] env[61995]: DEBUG oslo_concurrency.lockutils [req-f35bf19f-9547-4aef-b502-0f40e8f3ae48 req-28ad7baf-cbef-4cf1-9dab-8f30929a989c service nova] Releasing lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.423952] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Acquired lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.423952] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 645.476444] env[61995]: DEBUG nova.compute.utils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.481848] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.481848] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 645.522621] env[61995]: DEBUG nova.policy [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05da8959819d4bc49627e40f982f9523', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '156af3db43ac4ece9fea259c8dfa3ee6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 645.612633] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071619} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.612920] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 645.613785] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329a5e0a-27a0-4ee0-b3b3-337f55bf3352 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.636160] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 645.636160] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-314fec9f-93fc-4380-9a95-39e035de739f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.657164] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 645.657164] env[61995]: value = "task-794167" [ 645.657164] env[61995]: _type = "Task" [ 645.657164] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.665166] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794167, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.908324] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Successfully created port: 8f75df4e-c4cc-4102-86b9-61563de33e05 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.954711] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.982449] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.166409] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794167, 'name': ReconfigVM_Task, 'duration_secs': 0.367485} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.166764] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 646.170623] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.171816] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58ba914b-5628-4d45-b866-04c866484481 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.183422] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 646.183422] env[61995]: value = "task-794168" [ 646.183422] env[61995]: _type = "Task" [ 646.183422] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.197729] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794168, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.400943] env[61995]: DEBUG nova.compute.manager [req-eb800e59-8871-4f86-bdcf-c2139f0bd360 req-c82c2737-7c6c-4a43-9550-4429e8bad7d3 service nova] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Received event network-vif-deleted-9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 646.456209] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de380af2-2ab7-4e24-9f22-4cee19ed2c2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.464048] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1419299-5fca-44bd-ae17-eb34901cf585 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.500527] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be04962-9770-486a-8b5e-703e2f9de99a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.507977] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae420067-f1c9-4edb-a5e0-5d85e6598320 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.523261] env[61995]: DEBUG nova.compute.provider_tree [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.681252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Releasing lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.681252] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 646.681252] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 646.681252] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3fba01da-b511-4dec-9932-5640631af185 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.698764] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50f6c8d-ae49-437b-bcfd-d8acad56f108 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.718378] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794168, 'name': Rename_Task, 'duration_secs': 0.13337} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.718660] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 646.718913] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b846406-0cce-4b29-bc7a-1ee6ff92b203 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.724659] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance adb60754-9217-4b1b-b5cd-98553ff5bbbf could not be found. [ 646.724880] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 646.725038] env[61995]: INFO nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 646.725272] env[61995]: DEBUG oslo.service.loopingcall [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.726020] env[61995]: DEBUG nova.compute.manager [-] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.726020] env[61995]: DEBUG nova.network.neutron [-] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 646.730655] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 646.730655] env[61995]: value = "task-794169" [ 646.730655] env[61995]: _type = "Task" [ 646.730655] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.739953] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.742445] env[61995]: DEBUG nova.network.neutron [-] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.001570] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.029414] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:54:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1268571181',id=30,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1145596428',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.029644] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.031077] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.031077] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.031077] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.031077] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.031077] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.031267] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.031267] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.032142] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.032142] env[61995]: DEBUG nova.virt.hardware [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.032932] env[61995]: DEBUG nova.scheduler.client.report [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.037091] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e10686-f528-4481-9119-5ac3eb85a0e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.048189] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcb302d-46de-47e4-9f95-e6f9868c808c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.223510] env[61995]: ERROR nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. [ 647.223510] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.223510] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.223510] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.223510] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.223510] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.223510] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.223510] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.223510] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.223510] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 647.223510] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.223510] env[61995]: ERROR nova.compute.manager raise self.value [ 647.223510] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.223510] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.223510] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.223510] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.224122] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.224122] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.224122] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. [ 647.224122] env[61995]: ERROR nova.compute.manager [ 647.226063] env[61995]: Traceback (most recent call last): [ 647.226063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.226063] env[61995]: listener.cb(fileno) [ 647.226063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.226063] env[61995]: result = function(*args, **kwargs) [ 647.226063] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.226063] env[61995]: return func(*args, **kwargs) [ 647.226063] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.226063] env[61995]: raise e [ 647.226063] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.226063] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 647.226063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.226063] env[61995]: created_port_ids = self._update_ports_for_instance( [ 647.226063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.226063] env[61995]: with excutils.save_and_reraise_exception(): [ 647.226063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.226063] env[61995]: self.force_reraise() [ 647.226063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.226063] env[61995]: raise self.value [ 647.226063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.226063] env[61995]: updated_port = self._update_port( [ 647.226063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.226063] env[61995]: _ensure_no_port_binding_failure(port) [ 647.226063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.226063] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.226063] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. [ 647.226063] env[61995]: Removing descriptor: 17 [ 647.227198] env[61995]: ERROR nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Traceback (most recent call last): [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] yield resources [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self.driver.spawn(context, instance, image_meta, [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] vm_ref = self.build_virtual_machine(instance, [ 647.227198] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] for vif in network_info: [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return self._sync_wrapper(fn, *args, **kwargs) [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self.wait() [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self[:] = self._gt.wait() [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return self._exit_event.wait() [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.227590] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] result = hub.switch() [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return self.greenlet.switch() [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] result = function(*args, **kwargs) [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return func(*args, **kwargs) [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] raise e [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] nwinfo = self.network_api.allocate_for_instance( [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] created_port_ids = self._update_ports_for_instance( [ 647.228093] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] with excutils.save_and_reraise_exception(): [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self.force_reraise() [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] raise self.value [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] updated_port = self._update_port( [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] _ensure_no_port_binding_failure(port) [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] raise exception.PortBindingFailed(port_id=port['id']) [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] nova.exception.PortBindingFailed: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. [ 647.228526] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] [ 647.228973] env[61995]: INFO nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Terminating instance [ 647.234190] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Acquiring lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.234190] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Acquired lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.234190] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 647.245205] env[61995]: DEBUG oslo_vmware.api [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794169, 'name': PowerOnVM_Task, 'duration_secs': 0.420274} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.245925] env[61995]: DEBUG nova.network.neutron [-] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.247089] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 647.247137] env[61995]: INFO nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Took 7.57 seconds to spawn the instance on the hypervisor. [ 647.247738] env[61995]: DEBUG nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 647.249824] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7239f928-e75c-43ae-9172-4cd22ed3bcbd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.544207] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.544207] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.546590] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.663s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.547796] env[61995]: INFO nova.compute.claims [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.748445] env[61995]: INFO nova.compute.manager [-] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Took 1.02 seconds to deallocate network for instance. [ 647.752646] env[61995]: DEBUG nova.compute.claims [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 647.752857] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.758939] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.768599] env[61995]: INFO nova.compute.manager [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Took 37.29 seconds to build instance. [ 647.888602] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.056488] env[61995]: DEBUG nova.compute.utils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.060192] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.060192] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 648.146347] env[61995]: INFO nova.compute.manager [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Rebuilding instance [ 648.163813] env[61995]: DEBUG nova.policy [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '073aeaa337cd447894daee48142bb039', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb08165df7e04767839a0770ee8d73d2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 648.202530] env[61995]: DEBUG nova.compute.manager [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 648.203820] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b51e3d1-0928-4f67-88d5-a4481d9c823d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.273748] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9577d635-8062-4dc6-b546-bb9013f716a3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.245s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.390897] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Releasing lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.391483] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.391574] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 648.391876] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4dc2c40-86f8-4a91-8b9f-48372e4199e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.400938] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fff2e3f-d0d9-4d60-9553-619b7d0805bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.429631] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 596463d7-04ca-4f29-8adf-c451310aaf15 could not be found. [ 648.429853] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 648.430046] env[61995]: INFO nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Took 0.04 seconds to destroy the instance on the hypervisor. [ 648.430335] env[61995]: DEBUG oslo.service.loopingcall [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.432784] env[61995]: DEBUG nova.compute.manager [-] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.432883] env[61995]: DEBUG nova.network.neutron [-] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.435729] env[61995]: DEBUG nova.compute.manager [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Received event network-changed-8f75df4e-c4cc-4102-86b9-61563de33e05 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 648.435979] env[61995]: DEBUG nova.compute.manager [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Refreshing instance network info cache due to event network-changed-8f75df4e-c4cc-4102-86b9-61563de33e05. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 648.436208] env[61995]: DEBUG oslo_concurrency.lockutils [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] Acquiring lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.436349] env[61995]: DEBUG oslo_concurrency.lockutils [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] Acquired lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.436503] env[61995]: DEBUG nova.network.neutron [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Refreshing network info cache for port 8f75df4e-c4cc-4102-86b9-61563de33e05 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 648.470463] env[61995]: DEBUG nova.network.neutron [-] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.564709] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.720491] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 648.724597] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e2deb6a-b05f-4d58-b818-d039a4619b0f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.732329] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 648.732329] env[61995]: value = "task-794170" [ 648.732329] env[61995]: _type = "Task" [ 648.732329] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.744114] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.780996] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.966099] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Successfully created port: 40f80d5f-ab21-4529-8c1a-c8f59f288ea9 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.968387] env[61995]: DEBUG nova.network.neutron [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.975472] env[61995]: DEBUG nova.network.neutron [-] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.087323] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa84b47d-1177-4622-8a15-29d156371169 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.100574] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d668c1-6ced-432b-a21a-25ed4ec54d79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.140823] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d986e01e-803c-407c-ae2f-ff6b1ba0e640 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.148958] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f567c15d-e556-4162-a10f-1b2f6b808509 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.162800] env[61995]: DEBUG nova.compute.provider_tree [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.223107] env[61995]: DEBUG nova.network.neutron [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.242545] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794170, 'name': PowerOffVM_Task, 'duration_secs': 0.1841} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.242773] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 649.242984] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 649.243822] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bed8d2-9c56-4c1c-9dc6-c20d85556425 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.252016] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 649.252248] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12376689-66a4-4d7b-a055-72819401aa5a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.279483] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 649.279692] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 649.279938] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleting the datastore file [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 649.280121] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1db91e21-47d0-475b-99a1-08bd3854e30a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.288072] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 649.288072] env[61995]: value = "task-794172" [ 649.288072] env[61995]: _type = "Task" [ 649.288072] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.299133] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.313242] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.478499] env[61995]: INFO nova.compute.manager [-] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Took 1.05 seconds to deallocate network for instance. [ 649.483941] env[61995]: DEBUG nova.compute.claims [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.484148] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.573231] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.600675] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.600903] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.601067] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.601289] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.601464] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.601606] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.601807] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.601954] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.602147] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.602282] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.602445] env[61995]: DEBUG nova.virt.hardware [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.603338] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced0349a-8ce0-4c4e-affc-b24038521b81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.611580] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9be029-990e-4d58-ac50-104e2b9ac29c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.665424] env[61995]: DEBUG nova.scheduler.client.report [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.726015] env[61995]: DEBUG oslo_concurrency.lockutils [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] Releasing lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.726325] env[61995]: DEBUG nova.compute.manager [req-fd7bd1a9-5717-4bd0-8c21-14493fab6dc3 req-1af49eba-a46b-4167-9f30-a59756c90f87 service nova] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Received event network-vif-deleted-8f75df4e-c4cc-4102-86b9-61563de33e05 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 649.797707] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095444} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.798114] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 649.798365] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 649.798546] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 650.176215] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.627s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.176215] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.179115] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.102s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.181088] env[61995]: INFO nova.compute.claims [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.673178] env[61995]: DEBUG nova.compute.manager [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Received event network-changed-40f80d5f-ab21-4529-8c1a-c8f59f288ea9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 650.673178] env[61995]: DEBUG nova.compute.manager [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Refreshing instance network info cache due to event network-changed-40f80d5f-ab21-4529-8c1a-c8f59f288ea9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 650.673178] env[61995]: DEBUG oslo_concurrency.lockutils [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] Acquiring lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.673178] env[61995]: DEBUG oslo_concurrency.lockutils [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] Acquired lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.673178] env[61995]: DEBUG nova.network.neutron [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Refreshing network info cache for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 650.680918] env[61995]: DEBUG nova.compute.utils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.682973] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.684231] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 650.812727] env[61995]: DEBUG nova.policy [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7e1129f16ae479f9f77d03439cd1211', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb4715780da643999551e223caa16065', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 650.839812] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.840074] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.840253] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.840460] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.840607] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.841518] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.841518] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.841518] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.841518] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.841518] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.841993] env[61995]: DEBUG nova.virt.hardware [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.842464] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eff640-8f6c-4261-bf99-9fe40009d858 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.850897] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcda75e2-5958-4487-9ba2-4449574d3dc4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.864060] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 650.869553] env[61995]: DEBUG oslo.service.loopingcall [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.869796] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 650.869997] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d4ce4a1-7fa2-4460-b5f1-897cc3400bc8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.888838] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 650.888838] env[61995]: value = "task-794173" [ 650.888838] env[61995]: _type = "Task" [ 650.888838] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.896657] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794173, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.968493] env[61995]: ERROR nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. [ 650.968493] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 650.968493] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.968493] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 650.968493] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.968493] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 650.968493] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.968493] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 650.968493] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.968493] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 650.968493] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.968493] env[61995]: ERROR nova.compute.manager raise self.value [ 650.968493] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.968493] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 650.968493] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.968493] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 650.969063] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.969063] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 650.969063] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. [ 650.969063] env[61995]: ERROR nova.compute.manager [ 650.969063] env[61995]: Traceback (most recent call last): [ 650.969063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 650.969063] env[61995]: listener.cb(fileno) [ 650.969063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.969063] env[61995]: result = function(*args, **kwargs) [ 650.969063] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.969063] env[61995]: return func(*args, **kwargs) [ 650.969063] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.969063] env[61995]: raise e [ 650.969063] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.969063] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 650.969063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.969063] env[61995]: created_port_ids = self._update_ports_for_instance( [ 650.969063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.969063] env[61995]: with excutils.save_and_reraise_exception(): [ 650.969063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.969063] env[61995]: self.force_reraise() [ 650.969063] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.969063] env[61995]: raise self.value [ 650.969063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.969063] env[61995]: updated_port = self._update_port( [ 650.969063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.969063] env[61995]: _ensure_no_port_binding_failure(port) [ 650.969063] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.969063] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 650.969892] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. [ 650.969892] env[61995]: Removing descriptor: 17 [ 650.969892] env[61995]: ERROR nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Traceback (most recent call last): [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] yield resources [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self.driver.spawn(context, instance, image_meta, [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.969892] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] vm_ref = self.build_virtual_machine(instance, [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] for vif in network_info: [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return self._sync_wrapper(fn, *args, **kwargs) [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self.wait() [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self[:] = self._gt.wait() [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return self._exit_event.wait() [ 650.970261] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] result = hub.switch() [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return self.greenlet.switch() [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] result = function(*args, **kwargs) [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return func(*args, **kwargs) [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] raise e [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] nwinfo = self.network_api.allocate_for_instance( [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.970632] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] created_port_ids = self._update_ports_for_instance( [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] with excutils.save_and_reraise_exception(): [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self.force_reraise() [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] raise self.value [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] updated_port = self._update_port( [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] _ensure_no_port_binding_failure(port) [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.970999] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] raise exception.PortBindingFailed(port_id=port['id']) [ 650.971351] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] nova.exception.PortBindingFailed: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. [ 650.971351] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] [ 650.971351] env[61995]: INFO nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Terminating instance [ 650.972720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Acquiring lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.188117] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.206112] env[61995]: DEBUG nova.network.neutron [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.349833] env[61995]: DEBUG nova.network.neutron [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.397214] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794173, 'name': CreateVM_Task, 'duration_secs': 0.261789} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.399575] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 651.400421] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.400580] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.400948] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 651.401167] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6874ef4-05b2-4f62-9c31-d92e4292efe4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.405868] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 651.405868] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e3453f-145b-16c3-d43d-ec2f2bc01352" [ 651.405868] env[61995]: _type = "Task" [ 651.405868] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.414574] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e3453f-145b-16c3-d43d-ec2f2bc01352, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.574635] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Successfully created port: 020351b2-23a4-472f-96fd-d7ecf48067f0 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.614868] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f79d1b-485b-404e-b199-a6a724ab0b52 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.623767] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5718c7c-e57a-4984-ad57-155d49835b21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.657747] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c111d572-a0c9-45ea-8468-c7d70dcb467f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.669115] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ddabb7-7d50-4a9f-be95-be0579516d90 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.686200] env[61995]: DEBUG nova.compute.provider_tree [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.853958] env[61995]: DEBUG oslo_concurrency.lockutils [req-41344e65-9fde-41e8-87dc-134788f612cb req-c77acbf9-9c8f-42de-bd1c-4fe828405042 service nova] Releasing lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.854410] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Acquired lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.854595] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 651.917166] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e3453f-145b-16c3-d43d-ec2f2bc01352, 'name': SearchDatastore_Task, 'duration_secs': 0.008327} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.917631] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.918016] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 651.918412] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.918685] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.919053] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 651.919452] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-078d7834-4d9b-40d9-a266-67e198fbf16e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.927277] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 651.927524] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 651.928356] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbacf012-b0ba-438b-ae0f-4795b8f3cff1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.933769] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 651.933769] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5279ccc3-5dbf-0f75-77ee-feb20870eee3" [ 651.933769] env[61995]: _type = "Task" [ 651.933769] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.941408] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5279ccc3-5dbf-0f75-77ee-feb20870eee3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.189832] env[61995]: DEBUG nova.scheduler.client.report [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.198952] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.230703] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.230958] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.231080] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.231268] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.231417] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.231668] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.231880] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.232045] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.232216] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.232418] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.235816] env[61995]: DEBUG nova.virt.hardware [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.235816] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260c371a-b678-4d49-8845-c868310a4acc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.242631] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9748698e-28ef-4220-a7a6-a70c42a5ebce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.376146] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.445396] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5279ccc3-5dbf-0f75-77ee-feb20870eee3, 'name': SearchDatastore_Task, 'duration_secs': 0.00729} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.446153] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d379762-c8ae-450b-a0db-70a12ff1858e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.450997] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 652.450997] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52bba53a-1cac-910a-1b2f-940c4e7ca7d8" [ 652.450997] env[61995]: _type = "Task" [ 652.450997] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.458010] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52bba53a-1cac-910a-1b2f-940c4e7ca7d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.458747] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.696964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.697768] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 652.701060] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.707s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.702927] env[61995]: INFO nova.compute.claims [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.741279] env[61995]: DEBUG nova.compute.manager [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Received event network-vif-deleted-40f80d5f-ab21-4529-8c1a-c8f59f288ea9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 652.741621] env[61995]: DEBUG nova.compute.manager [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Received event network-changed-020351b2-23a4-472f-96fd-d7ecf48067f0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 652.741621] env[61995]: DEBUG nova.compute.manager [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Refreshing instance network info cache due to event network-changed-020351b2-23a4-472f-96fd-d7ecf48067f0. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 652.741867] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] Acquiring lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.741867] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] Acquired lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.742020] env[61995]: DEBUG nova.network.neutron [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Refreshing network info cache for port 020351b2-23a4-472f-96fd-d7ecf48067f0 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 652.871982] env[61995]: ERROR nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. [ 652.871982] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.871982] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.871982] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.871982] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.871982] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.871982] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.871982] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.871982] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.871982] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 652.871982] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.871982] env[61995]: ERROR nova.compute.manager raise self.value [ 652.871982] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.871982] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.871982] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.871982] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.872508] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.872508] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.872508] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. [ 652.872508] env[61995]: ERROR nova.compute.manager [ 652.872508] env[61995]: Traceback (most recent call last): [ 652.872508] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.872508] env[61995]: listener.cb(fileno) [ 652.872508] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.872508] env[61995]: result = function(*args, **kwargs) [ 652.872508] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.872508] env[61995]: return func(*args, **kwargs) [ 652.872508] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.872508] env[61995]: raise e [ 652.872508] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.872508] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 652.872508] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.872508] env[61995]: created_port_ids = self._update_ports_for_instance( [ 652.872508] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.872508] env[61995]: with excutils.save_and_reraise_exception(): [ 652.872508] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.872508] env[61995]: self.force_reraise() [ 652.872508] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.872508] env[61995]: raise self.value [ 652.872508] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.872508] env[61995]: updated_port = self._update_port( [ 652.872508] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.872508] env[61995]: _ensure_no_port_binding_failure(port) [ 652.872508] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.872508] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.873642] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. [ 652.873642] env[61995]: Removing descriptor: 14 [ 652.873642] env[61995]: ERROR nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Traceback (most recent call last): [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] yield resources [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self.driver.spawn(context, instance, image_meta, [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.873642] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] vm_ref = self.build_virtual_machine(instance, [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] for vif in network_info: [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return self._sync_wrapper(fn, *args, **kwargs) [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self.wait() [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self[:] = self._gt.wait() [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return self._exit_event.wait() [ 652.874079] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] result = hub.switch() [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return self.greenlet.switch() [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] result = function(*args, **kwargs) [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return func(*args, **kwargs) [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] raise e [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] nwinfo = self.network_api.allocate_for_instance( [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.874466] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] created_port_ids = self._update_ports_for_instance( [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] with excutils.save_and_reraise_exception(): [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self.force_reraise() [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] raise self.value [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] updated_port = self._update_port( [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] _ensure_no_port_binding_failure(port) [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.875132] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] raise exception.PortBindingFailed(port_id=port['id']) [ 652.875644] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] nova.exception.PortBindingFailed: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. [ 652.875644] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] [ 652.875644] env[61995]: INFO nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Terminating instance [ 652.876475] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.962415] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Releasing lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.962840] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.963046] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 652.963348] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52bba53a-1cac-910a-1b2f-940c4e7ca7d8, 'name': SearchDatastore_Task, 'duration_secs': 0.008854} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.963552] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15e917f6-a9cf-4c72-b5cd-1aa529e768f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.965474] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.965813] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 652.966262] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30451116-9f52-4a1b-8c35-32feb6985b69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.975662] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57dfbe1-a5c5-4c67-b8ee-42be4af319a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.986925] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 652.986925] env[61995]: value = "task-794174" [ 652.986925] env[61995]: _type = "Task" [ 652.986925] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.996137] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794174, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.001206] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3252888a-31ec-4d93-a8e2-b67aa0bab1e8 could not be found. [ 653.001584] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 653.001791] env[61995]: INFO nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 653.002067] env[61995]: DEBUG oslo.service.loopingcall [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.002329] env[61995]: DEBUG nova.compute.manager [-] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.002329] env[61995]: DEBUG nova.network.neutron [-] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.018868] env[61995]: DEBUG nova.network.neutron [-] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.207491] env[61995]: DEBUG nova.compute.utils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.213106] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.213106] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 653.266273] env[61995]: DEBUG nova.network.neutron [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.286132] env[61995]: DEBUG nova.policy [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d83037b65b74c7bb8d0ddaa02c2dc4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5320229bdb8143598497840a99b5d0d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 653.441944] env[61995]: DEBUG nova.network.neutron [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.498414] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794174, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434185} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.498414] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 653.498414] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 653.498414] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25e84c74-dd5e-4c57-8713-20f3b367f945 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.506392] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 653.506392] env[61995]: value = "task-794175" [ 653.506392] env[61995]: _type = "Task" [ 653.506392] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.515236] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794175, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.520957] env[61995]: DEBUG nova.network.neutron [-] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.669716] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Successfully created port: f761173d-e9c8-4514-9ee8-2ad47f90a6b3 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.713593] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 653.945843] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e326bce-27c9-4c8d-a77c-5e027f523c8b req-63aa751c-9e21-470a-bd3d-edc980770496 service nova] Releasing lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.947484] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquired lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.947773] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.021733] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794175, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061041} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.021999] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 654.023759] env[61995]: INFO nova.compute.manager [-] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Took 1.02 seconds to deallocate network for instance. [ 654.024840] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfd2a93-ecf7-4030-90d3-cda9269366d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.031406] env[61995]: DEBUG nova.compute.claims [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.031646] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.047682] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 654.050187] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-005d73b7-ecde-4cc4-b0f8-c0a0caf815cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.068579] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 654.068579] env[61995]: value = "task-794176" [ 654.068579] env[61995]: _type = "Task" [ 654.068579] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.076448] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794176, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.167268] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1308bbf-725c-435b-9156-8c8da5c8b83e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.174904] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63002727-6afc-48df-bd33-16a2e814edb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.209950] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864947cf-ec7a-4695-9bd3-2e3bd12f21bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.217997] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0563c0b-4157-410c-9013-68a7da05e501 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.236290] env[61995]: DEBUG nova.compute.provider_tree [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.467384] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.579135] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794176, 'name': ReconfigVM_Task, 'duration_secs': 0.257952} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.579440] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 654.580052] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6999d9a-37bb-4067-9bb5-cf38d323e8af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.586320] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 654.586320] env[61995]: value = "task-794177" [ 654.586320] env[61995]: _type = "Task" [ 654.586320] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.594210] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794177, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.603958] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.738867] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.742727] env[61995]: DEBUG nova.scheduler.client.report [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.771863] env[61995]: DEBUG nova.compute.manager [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Received event network-vif-deleted-020351b2-23a4-472f-96fd-d7ecf48067f0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 654.771863] env[61995]: DEBUG nova.compute.manager [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Received event network-changed-f761173d-e9c8-4514-9ee8-2ad47f90a6b3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 654.771863] env[61995]: DEBUG nova.compute.manager [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Refreshing instance network info cache due to event network-changed-f761173d-e9c8-4514-9ee8-2ad47f90a6b3. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 654.771990] env[61995]: DEBUG oslo_concurrency.lockutils [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] Acquiring lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.772219] env[61995]: DEBUG oslo_concurrency.lockutils [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] Acquired lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.772262] env[61995]: DEBUG nova.network.neutron [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Refreshing network info cache for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 654.775972] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.776233] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.776390] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.776570] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.776824] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.776904] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.777069] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.777233] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.777397] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.777557] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.777722] env[61995]: DEBUG nova.virt.hardware [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.779366] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a39998a-9cdc-49f4-abfa-4e8f1d4b2df9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.790439] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2166b08f-8ea3-4d94-a881-b86f0dc886f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.829519] env[61995]: DEBUG nova.network.neutron [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.935231] env[61995]: ERROR nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. [ 654.935231] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.935231] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.935231] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.935231] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.935231] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.935231] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.935231] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.935231] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.935231] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 654.935231] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.935231] env[61995]: ERROR nova.compute.manager raise self.value [ 654.935231] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.935231] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.935231] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.935231] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.935790] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.935790] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.935790] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. [ 654.935790] env[61995]: ERROR nova.compute.manager [ 654.935790] env[61995]: Traceback (most recent call last): [ 654.935790] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.935790] env[61995]: listener.cb(fileno) [ 654.935790] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.935790] env[61995]: result = function(*args, **kwargs) [ 654.935790] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.935790] env[61995]: return func(*args, **kwargs) [ 654.935790] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.935790] env[61995]: raise e [ 654.935790] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.935790] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 654.935790] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.935790] env[61995]: created_port_ids = self._update_ports_for_instance( [ 654.935790] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.935790] env[61995]: with excutils.save_and_reraise_exception(): [ 654.935790] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.935790] env[61995]: self.force_reraise() [ 654.935790] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.935790] env[61995]: raise self.value [ 654.935790] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.935790] env[61995]: updated_port = self._update_port( [ 654.935790] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.935790] env[61995]: _ensure_no_port_binding_failure(port) [ 654.935790] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.935790] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.936751] env[61995]: nova.exception.PortBindingFailed: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. [ 654.936751] env[61995]: Removing descriptor: 14 [ 654.936751] env[61995]: ERROR nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Traceback (most recent call last): [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] yield resources [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self.driver.spawn(context, instance, image_meta, [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.936751] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] vm_ref = self.build_virtual_machine(instance, [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] for vif in network_info: [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return self._sync_wrapper(fn, *args, **kwargs) [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self.wait() [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self[:] = self._gt.wait() [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return self._exit_event.wait() [ 654.937160] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] result = hub.switch() [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return self.greenlet.switch() [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] result = function(*args, **kwargs) [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return func(*args, **kwargs) [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] raise e [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] nwinfo = self.network_api.allocate_for_instance( [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.937593] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] created_port_ids = self._update_ports_for_instance( [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] with excutils.save_and_reraise_exception(): [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self.force_reraise() [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] raise self.value [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] updated_port = self._update_port( [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] _ensure_no_port_binding_failure(port) [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.938626] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] raise exception.PortBindingFailed(port_id=port['id']) [ 654.939473] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] nova.exception.PortBindingFailed: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. [ 654.939473] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] [ 654.939473] env[61995]: INFO nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Terminating instance [ 654.939473] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Acquiring lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.969693] env[61995]: DEBUG nova.network.neutron [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.096036] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794177, 'name': Rename_Task, 'duration_secs': 0.12916} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.096289] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 655.096523] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cceb8664-9b81-40cb-9056-002faf1b03ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.102368] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 655.102368] env[61995]: value = "task-794178" [ 655.102368] env[61995]: _type = "Task" [ 655.102368] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.106856] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Releasing lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.107256] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.107432] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 655.108192] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7e86b1f-b89e-427c-a9b4-4eb1314b9bb4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.113656] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.117163] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494abe01-a981-4273-bf0b-0c8ba0c1f914 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.142338] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 862694a5-63f8-4e2d-982a-329f4a071d29 could not be found. [ 655.142584] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.142765] env[61995]: INFO nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.143018] env[61995]: DEBUG oslo.service.loopingcall [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.143250] env[61995]: DEBUG nova.compute.manager [-] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.143348] env[61995]: DEBUG nova.network.neutron [-] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.158558] env[61995]: DEBUG nova.network.neutron [-] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.247774] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.248461] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.251248] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.008s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.472094] env[61995]: DEBUG oslo_concurrency.lockutils [req-8a81adb8-164c-4f59-8079-3f3007c682a5 req-21c66b52-0dc3-4a27-a36b-938b62b34924 service nova] Releasing lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.472568] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Acquired lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.472758] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 655.612337] env[61995]: DEBUG oslo_vmware.api [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794178, 'name': PowerOnVM_Task, 'duration_secs': 0.436434} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.612672] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 655.612918] env[61995]: DEBUG nova.compute.manager [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 655.613707] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a959f4c-63e1-4f57-b5f9-f4cbf023724e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.661357] env[61995]: DEBUG nova.network.neutron [-] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.756865] env[61995]: DEBUG nova.compute.utils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.762295] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.762295] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 655.801377] env[61995]: DEBUG nova.policy [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7114a95a50664a97bc1a3978341c63b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2a43731912c441782c38522c2187c8f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 655.962595] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.962833] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.007558] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.131494] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.136843] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934e568c-332f-47fc-b91f-e4ffae9ce337 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.144016] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b39137f-b41b-485b-afce-29ef398c7a47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.147861] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.180965] env[61995]: INFO nova.compute.manager [-] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Took 1.04 seconds to deallocate network for instance. [ 656.183838] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61390326-9918-4839-9f9b-37f555396bce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.186587] env[61995]: DEBUG nova.compute.claims [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.187913] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.192274] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a947672a-20d2-493c-be12-e93a30f97fc9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.205636] env[61995]: DEBUG nova.compute.provider_tree [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.263090] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.380670] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Successfully created port: d861c7f2-f048-4f98-b1a3-61ac952502d2 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.654054] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Releasing lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.654054] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 656.654054] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 656.654054] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acf5d25e-3c63-4311-a04a-ba6449682cb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.661092] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48e53a6-b47d-4aed-87d1-8cf3082ded02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.683783] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7bdf0b12-25d2-45a5-af57-b1bb31148f3d could not be found. [ 656.684171] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 656.684503] env[61995]: INFO nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 656.684882] env[61995]: DEBUG oslo.service.loopingcall [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.685225] env[61995]: DEBUG nova.compute.manager [-] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.685454] env[61995]: DEBUG nova.network.neutron [-] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 656.711019] env[61995]: DEBUG nova.scheduler.client.report [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.715992] env[61995]: DEBUG nova.network.neutron [-] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.859274] env[61995]: DEBUG nova.compute.manager [req-3c4f3927-e53e-4bad-b175-37527ace3ff9 req-1788093c-b845-42d9-9734-f216820d6684 service nova] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Received event network-vif-deleted-f761173d-e9c8-4514-9ee8-2ad47f90a6b3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 657.218364] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.218364] env[61995]: ERROR nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. [ 657.218364] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Traceback (most recent call last): [ 657.218364] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.218364] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self.driver.spawn(context, instance, image_meta, [ 657.218364] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.218364] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.218364] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.218364] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] vm_ref = self.build_virtual_machine(instance, [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] for vif in network_info: [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return self._sync_wrapper(fn, *args, **kwargs) [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self.wait() [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self[:] = self._gt.wait() [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return self._exit_event.wait() [ 657.218990] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] result = hub.switch() [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return self.greenlet.switch() [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] result = function(*args, **kwargs) [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] return func(*args, **kwargs) [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] raise e [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] nwinfo = self.network_api.allocate_for_instance( [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.219388] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] created_port_ids = self._update_ports_for_instance( [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] with excutils.save_and_reraise_exception(): [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] self.force_reraise() [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] raise self.value [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] updated_port = self._update_port( [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] _ensure_no_port_binding_failure(port) [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.219764] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] raise exception.PortBindingFailed(port_id=port['id']) [ 657.220128] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] nova.exception.PortBindingFailed: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. [ 657.220128] env[61995]: ERROR nova.compute.manager [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] [ 657.220128] env[61995]: DEBUG nova.compute.utils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.222542] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.858s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.225549] env[61995]: DEBUG nova.network.neutron [-] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.226804] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Build of instance ac219077-ed13-45ee-8d47-ff5dd5abe89a was re-scheduled: Binding failed for port 5c8b7665-348d-4130-bf59-fbf14180b7e0, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.227313] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.227644] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Acquiring lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.227862] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Acquired lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.228114] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.273154] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.296612] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.296719] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.296939] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.297204] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.297341] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.297489] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.297694] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.297848] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.298131] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.298219] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.298377] env[61995]: DEBUG nova.virt.hardware [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.299458] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93993ac-6aa1-4562-9741-61d95b2ef7a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.307300] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513b0deb-140d-4d4f-9b34-0f8e0de609e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.442811] env[61995]: ERROR nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. [ 657.442811] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.442811] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.442811] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.442811] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.442811] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.442811] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.442811] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.442811] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.442811] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 657.442811] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.442811] env[61995]: ERROR nova.compute.manager raise self.value [ 657.442811] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.442811] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.442811] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.442811] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.443366] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.443366] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.443366] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. [ 657.443366] env[61995]: ERROR nova.compute.manager [ 657.443366] env[61995]: Traceback (most recent call last): [ 657.443366] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.443366] env[61995]: listener.cb(fileno) [ 657.443366] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.443366] env[61995]: result = function(*args, **kwargs) [ 657.443366] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.443366] env[61995]: return func(*args, **kwargs) [ 657.443366] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.443366] env[61995]: raise e [ 657.443366] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.443366] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 657.443366] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.443366] env[61995]: created_port_ids = self._update_ports_for_instance( [ 657.443366] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.443366] env[61995]: with excutils.save_and_reraise_exception(): [ 657.443366] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.443366] env[61995]: self.force_reraise() [ 657.443366] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.443366] env[61995]: raise self.value [ 657.443366] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.443366] env[61995]: updated_port = self._update_port( [ 657.443366] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.443366] env[61995]: _ensure_no_port_binding_failure(port) [ 657.443366] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.443366] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.444305] env[61995]: nova.exception.PortBindingFailed: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. [ 657.444305] env[61995]: Removing descriptor: 17 [ 657.444305] env[61995]: ERROR nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Traceback (most recent call last): [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] yield resources [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self.driver.spawn(context, instance, image_meta, [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.444305] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] vm_ref = self.build_virtual_machine(instance, [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] for vif in network_info: [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return self._sync_wrapper(fn, *args, **kwargs) [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self.wait() [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self[:] = self._gt.wait() [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return self._exit_event.wait() [ 657.444799] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] result = hub.switch() [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return self.greenlet.switch() [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] result = function(*args, **kwargs) [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return func(*args, **kwargs) [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] raise e [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] nwinfo = self.network_api.allocate_for_instance( [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.445259] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] created_port_ids = self._update_ports_for_instance( [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] with excutils.save_and_reraise_exception(): [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self.force_reraise() [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] raise self.value [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] updated_port = self._update_port( [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] _ensure_no_port_binding_failure(port) [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.445701] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] raise exception.PortBindingFailed(port_id=port['id']) [ 657.446424] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] nova.exception.PortBindingFailed: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. [ 657.446424] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] [ 657.446424] env[61995]: INFO nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Terminating instance [ 657.446734] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Acquiring lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.447016] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Acquired lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.447246] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.555981] env[61995]: INFO nova.compute.manager [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Rebuilding instance [ 657.600778] env[61995]: DEBUG nova.compute.manager [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 657.601670] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b2fa22-7ac9-49c6-9dc9-c2ec6c091b36 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.731124] env[61995]: INFO nova.compute.manager [-] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Took 1.05 seconds to deallocate network for instance. [ 657.735504] env[61995]: DEBUG nova.compute.claims [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 657.735597] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.746473] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.821481] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.966182] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.058697] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.083646] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a3751f-8d4d-4097-b1cd-895e9b615c6a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.091697] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac36d00-fbc6-4043-8073-029a167ba955 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.123302] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 658.123806] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b24c614f-ba57-40dd-a5f4-dea917439c4d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.126067] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7f8ac4-eb76-4801-b24a-75fb6b61190c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.135246] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cff8853-e5d5-4013-a7f8-89fdd07fd4ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.140190] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 658.140190] env[61995]: value = "task-794179" [ 658.140190] env[61995]: _type = "Task" [ 658.140190] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.151023] env[61995]: DEBUG nova.compute.provider_tree [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.156983] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.324226] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Releasing lock "refresh_cache-ac219077-ed13-45ee-8d47-ff5dd5abe89a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.325028] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.325028] env[61995]: DEBUG nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.325028] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.339246] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.566151] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Releasing lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.566151] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.566151] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 658.566151] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f1732b7-f75f-446f-882b-b0c8ac418ae8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.573623] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f12f9e-aa00-4862-af94-b02434ee85dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.596097] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ca64ce9-9856-4e65-a2e7-af11ae0211da could not be found. [ 658.596264] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 658.596443] env[61995]: INFO nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Took 0.03 seconds to destroy the instance on the hypervisor. [ 658.596684] env[61995]: DEBUG oslo.service.loopingcall [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.596903] env[61995]: DEBUG nova.compute.manager [-] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.596999] env[61995]: DEBUG nova.network.neutron [-] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.617478] env[61995]: DEBUG nova.network.neutron [-] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.650316] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794179, 'name': PowerOffVM_Task, 'duration_secs': 0.121852} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.650594] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 658.650808] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 658.651697] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0524e17-9861-4aab-92bd-1b7bd0a7360f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.654916] env[61995]: DEBUG nova.scheduler.client.report [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.662361] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 658.662792] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c98bf567-911f-4f6b-a89b-9c0cb7858b85 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.688395] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 658.688697] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 658.688958] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Deleting the datastore file [datastore2] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 658.689255] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06363a89-d62a-4be1-a23d-eef918552bd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.696270] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 658.696270] env[61995]: value = "task-794181" [ 658.696270] env[61995]: _type = "Task" [ 658.696270] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.704496] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794181, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.842329] env[61995]: DEBUG nova.network.neutron [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.889147] env[61995]: DEBUG nova.compute.manager [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Received event network-changed-d861c7f2-f048-4f98-b1a3-61ac952502d2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 658.889359] env[61995]: DEBUG nova.compute.manager [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Refreshing instance network info cache due to event network-changed-d861c7f2-f048-4f98-b1a3-61ac952502d2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 658.889567] env[61995]: DEBUG oslo_concurrency.lockutils [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] Acquiring lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.889706] env[61995]: DEBUG oslo_concurrency.lockutils [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] Acquired lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.889859] env[61995]: DEBUG nova.network.neutron [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Refreshing network info cache for port d861c7f2-f048-4f98-b1a3-61ac952502d2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 659.120133] env[61995]: DEBUG nova.network.neutron [-] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.163751] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.164456] env[61995]: ERROR nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Traceback (most recent call last): [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self.driver.spawn(context, instance, image_meta, [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] vm_ref = self.build_virtual_machine(instance, [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.164456] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] for vif in network_info: [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return self._sync_wrapper(fn, *args, **kwargs) [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self.wait() [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self[:] = self._gt.wait() [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return self._exit_event.wait() [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] result = hub.switch() [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 659.164747] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return self.greenlet.switch() [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] result = function(*args, **kwargs) [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] return func(*args, **kwargs) [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] raise e [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] nwinfo = self.network_api.allocate_for_instance( [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] created_port_ids = self._update_ports_for_instance( [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] with excutils.save_and_reraise_exception(): [ 659.165066] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] self.force_reraise() [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] raise self.value [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] updated_port = self._update_port( [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] _ensure_no_port_binding_failure(port) [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] raise exception.PortBindingFailed(port_id=port['id']) [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] nova.exception.PortBindingFailed: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. [ 659.165378] env[61995]: ERROR nova.compute.manager [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] [ 659.165641] env[61995]: DEBUG nova.compute.utils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.166286] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.649s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.167778] env[61995]: INFO nova.compute.claims [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.170692] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Build of instance 278cba4c-5bfb-42e0-b502-8f524ba1f380 was re-scheduled: Binding failed for port 80ba18c9-8004-46b7-ae1a-fc17685d183e, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.171109] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.171327] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Acquiring lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.171467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Acquired lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.171622] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 659.206391] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09984} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.206623] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 659.206797] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 659.206968] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 659.345198] env[61995]: INFO nova.compute.manager [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] [instance: ac219077-ed13-45ee-8d47-ff5dd5abe89a] Took 1.02 seconds to deallocate network for instance. [ 659.407044] env[61995]: DEBUG nova.network.neutron [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.496942] env[61995]: DEBUG nova.network.neutron [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.623526] env[61995]: INFO nova.compute.manager [-] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Took 1.03 seconds to deallocate network for instance. [ 659.625786] env[61995]: DEBUG nova.compute.claims [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.625997] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.698207] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.749472] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.999569] env[61995]: DEBUG oslo_concurrency.lockutils [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] Releasing lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.999845] env[61995]: DEBUG nova.compute.manager [req-d2065518-c512-4a07-8b56-427256f24d1e req-785d10b3-a2fd-4362-a5b3-c367e5dfd5b9 service nova] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Received event network-vif-deleted-d861c7f2-f048-4f98-b1a3-61ac952502d2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 660.238291] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.238591] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.238727] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.238860] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.239019] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.239191] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.239405] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.239560] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.239719] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.239878] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.240054] env[61995]: DEBUG nova.virt.hardware [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.240914] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c6fe63-0acc-48b6-b603-c47ffa2553d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.250863] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b729302f-cb67-44f3-80ff-1ca14e901b63 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.254860] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Releasing lock "refresh_cache-278cba4c-5bfb-42e0-b502-8f524ba1f380" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.255073] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 660.255240] env[61995]: DEBUG nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.255478] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 660.266156] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 660.271651] env[61995]: DEBUG oslo.service.loopingcall [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.274037] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 660.274423] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fee32ac5-7b50-4add-be0d-dceead1a5cb4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.288817] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.295032] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 660.295032] env[61995]: value = "task-794182" [ 660.295032] env[61995]: _type = "Task" [ 660.295032] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.306572] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794182, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.375310] env[61995]: INFO nova.scheduler.client.report [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Deleted allocations for instance ac219077-ed13-45ee-8d47-ff5dd5abe89a [ 660.584434] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ba36a1-d89e-4bab-b160-80761f7dd9c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.592997] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d2920c-a88b-412c-a56b-01e1b2607dd2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.624352] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fa1987-3a15-4738-bcf4-f15f11f7d1a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.631803] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3a4c51-5ad3-40a6-8e5d-49e1e15108ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.644554] env[61995]: DEBUG nova.compute.provider_tree [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.791423] env[61995]: DEBUG nova.network.neutron [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.805267] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794182, 'name': CreateVM_Task, 'duration_secs': 0.317085} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.805479] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 660.806322] env[61995]: DEBUG oslo_vmware.service [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067b14ed-7f48-41a5-896a-927e2d36c7bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.812048] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.812221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.812559] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 660.812786] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a15efe21-1401-4e46-9c6d-8dfd3b969da9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.816660] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 660.816660] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d8ca13-c08f-5e19-4859-62ef3fe876f1" [ 660.816660] env[61995]: _type = "Task" [ 660.816660] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.825110] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d8ca13-c08f-5e19-4859-62ef3fe876f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.886817] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8d9c732d-22df-426e-a0c3-fda9c50293ef tempest-ServerRescueTestJSONUnderV235-2144498279 tempest-ServerRescueTestJSONUnderV235-2144498279-project-member] Lock "ac219077-ed13-45ee-8d47-ff5dd5abe89a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.626s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.148200] env[61995]: DEBUG nova.scheduler.client.report [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.293991] env[61995]: INFO nova.compute.manager [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] [instance: 278cba4c-5bfb-42e0-b502-8f524ba1f380] Took 1.04 seconds to deallocate network for instance. [ 661.326908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.327192] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 661.327421] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.327564] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.327858] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 661.328059] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e28d8518-d0ba-487d-9865-ce122354d9e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.336904] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 661.336904] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 661.337677] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d05b7c2-36b8-419d-8864-8199f97c7395 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.343702] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eba5ec34-ab48-429f-a437-3533636740fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.348921] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 661.348921] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a74123-2071-22fb-a990-cb5ef7a59d35" [ 661.348921] env[61995]: _type = "Task" [ 661.348921] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.356236] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a74123-2071-22fb-a990-cb5ef7a59d35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.389722] env[61995]: DEBUG nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.653633] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.654189] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.657147] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.904s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.859616] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Preparing fetch location {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 661.859827] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Creating directory with path [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 661.860107] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-770870c0-b7d4-47cb-99d1-988289e19ea0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.871883] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Created directory with path [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 661.872119] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Fetch image to [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 661.872297] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Downloading image file data c1739132-aff3-4644-9522-fc02cf765bc8 to [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk on the data store datastore1 {{(pid=61995) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 661.873050] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81772422-56b9-470c-966e-5c8667425f06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.879893] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8508f36-2818-4f87-95ea-2c49426db624 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.890840] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4e3b92-ae80-4555-a268-9fe19494ce31 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.928383] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72dd2d2b-0698-40a0-ab6d-14960f76d852 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.933731] env[61995]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-0df9cdf7-cde8-44a0-b4cd-e02598b44bcf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.943111] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.956825] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Downloading image file data c1739132-aff3-4644-9522-fc02cf765bc8 to the data store datastore1 {{(pid=61995) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 662.008127] env[61995]: DEBUG oslo_vmware.rw_handles [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61995) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 662.166892] env[61995]: DEBUG nova.compute.utils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.171720] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.171891] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 662.253372] env[61995]: DEBUG nova.policy [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e49ca6cd1fe04c0a8e6e2e9945940798', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba083ad7c24c4aceba2cf151ad528d82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 662.326467] env[61995]: INFO nova.scheduler.client.report [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Deleted allocations for instance 278cba4c-5bfb-42e0-b502-8f524ba1f380 [ 662.633180] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db204f95-23f7-4248-a3eb-2ec436bc213b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.644716] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d67e179-4922-4456-8a91-8e807ff4fa32 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.648545] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Successfully created port: 328e9ca5-d5be-4008-a06c-74825b4d0e79 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.679515] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.683585] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3b0f24-311b-4669-8815-eeba313cc6ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.689450] env[61995]: DEBUG oslo_vmware.rw_handles [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Completed reading data from the image iterator. {{(pid=61995) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 662.690077] env[61995]: DEBUG oslo_vmware.rw_handles [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 662.695868] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d339b17e-4309-4e8d-bf2f-d7b7d4c33a55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.709943] env[61995]: DEBUG nova.compute.provider_tree [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.812591] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Downloaded image file data c1739132-aff3-4644-9522-fc02cf765bc8 to vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk on the data store datastore1 {{(pid=61995) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 662.814871] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Caching image {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 662.817146] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Copying Virtual Disk [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk to [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 662.817146] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0eb47c26-81d1-465e-b626-c48ef9990395 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.826536] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 662.826536] env[61995]: value = "task-794183" [ 662.826536] env[61995]: _type = "Task" [ 662.826536] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.840536] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92c286bc-0fc4-493d-88eb-fa9615a86bc6 tempest-TenantUsagesTestJSON-1723800201 tempest-TenantUsagesTestJSON-1723800201-project-member] Lock "278cba4c-5bfb-42e0-b502-8f524ba1f380" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.088s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.840536] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.215053] env[61995]: DEBUG nova.scheduler.client.report [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.337321] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794183, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.345039] env[61995]: DEBUG nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.698321] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.720119] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.062s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.720383] env[61995]: ERROR nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Traceback (most recent call last): [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self.driver.spawn(context, instance, image_meta, [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] vm_ref = self.build_virtual_machine(instance, [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.720383] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] for vif in network_info: [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] return self._sync_wrapper(fn, *args, **kwargs) [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self.wait() [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self[:] = self._gt.wait() [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] return self._exit_event.wait() [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] current.throw(*self._exc) [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.720641] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] result = function(*args, **kwargs) [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] return func(*args, **kwargs) [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] raise e [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] nwinfo = self.network_api.allocate_for_instance( [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] created_port_ids = self._update_ports_for_instance( [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] with excutils.save_and_reraise_exception(): [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] self.force_reraise() [ 663.720959] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] raise self.value [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] updated_port = self._update_port( [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] _ensure_no_port_binding_failure(port) [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] raise exception.PortBindingFailed(port_id=port['id']) [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] nova.exception.PortBindingFailed: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. [ 663.721308] env[61995]: ERROR nova.compute.manager [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] [ 663.721308] env[61995]: DEBUG nova.compute.utils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.724505] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.411s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.726032] env[61995]: INFO nova.compute.claims [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.729255] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Build of instance adb60754-9217-4b1b-b5cd-98553ff5bbbf was re-scheduled: Binding failed for port 9c511ea9-71b5-4b7a-bfe7-15ff74a5feb4, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.729940] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.730199] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Acquiring lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.730348] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Acquired lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.730503] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.745000] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.745277] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.745428] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.745605] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.745910] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.745910] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.746955] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.747207] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.747361] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.747526] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.747693] env[61995]: DEBUG nova.virt.hardware [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.748862] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d9d830-8baa-4859-88b1-4c9ce0492161 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.764853] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45745bb-5380-4784-b27b-fbc1a09784b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.771744] env[61995]: DEBUG nova.compute.manager [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Received event network-changed-328e9ca5-d5be-4008-a06c-74825b4d0e79 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 663.772043] env[61995]: DEBUG nova.compute.manager [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Refreshing instance network info cache due to event network-changed-328e9ca5-d5be-4008-a06c-74825b4d0e79. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 663.772337] env[61995]: DEBUG oslo_concurrency.lockutils [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] Acquiring lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.772588] env[61995]: DEBUG oslo_concurrency.lockutils [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] Acquired lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.773433] env[61995]: DEBUG nova.network.neutron [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Refreshing network info cache for port 328e9ca5-d5be-4008-a06c-74825b4d0e79 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 663.838180] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794183, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644215} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.838442] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Copied Virtual Disk [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk to [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 663.838616] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Deleting the datastore file [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8/tmp-sparse.vmdk {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 663.838850] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56349e0b-279c-461f-a4c5-ce980e969339 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.845183] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 663.845183] env[61995]: value = "task-794184" [ 663.845183] env[61995]: _type = "Task" [ 663.845183] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.855687] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794184, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.869987] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.953109] env[61995]: ERROR nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. [ 663.953109] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.953109] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.953109] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.953109] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.953109] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.953109] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.953109] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.953109] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.953109] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 663.953109] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.953109] env[61995]: ERROR nova.compute.manager raise self.value [ 663.953109] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.953109] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.953109] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.953109] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.953611] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.953611] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.953611] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. [ 663.953611] env[61995]: ERROR nova.compute.manager [ 663.953611] env[61995]: Traceback (most recent call last): [ 663.953611] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.953611] env[61995]: listener.cb(fileno) [ 663.953611] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.953611] env[61995]: result = function(*args, **kwargs) [ 663.953611] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.953611] env[61995]: return func(*args, **kwargs) [ 663.953611] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.953611] env[61995]: raise e [ 663.953611] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.953611] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 663.953611] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.953611] env[61995]: created_port_ids = self._update_ports_for_instance( [ 663.953611] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.953611] env[61995]: with excutils.save_and_reraise_exception(): [ 663.953611] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.953611] env[61995]: self.force_reraise() [ 663.953611] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.953611] env[61995]: raise self.value [ 663.953611] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.953611] env[61995]: updated_port = self._update_port( [ 663.953611] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.953611] env[61995]: _ensure_no_port_binding_failure(port) [ 663.953611] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.953611] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.954421] env[61995]: nova.exception.PortBindingFailed: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. [ 663.954421] env[61995]: Removing descriptor: 17 [ 663.954421] env[61995]: ERROR nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Traceback (most recent call last): [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] yield resources [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self.driver.spawn(context, instance, image_meta, [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.954421] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] vm_ref = self.build_virtual_machine(instance, [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] for vif in network_info: [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return self._sync_wrapper(fn, *args, **kwargs) [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self.wait() [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self[:] = self._gt.wait() [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return self._exit_event.wait() [ 663.954859] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] result = hub.switch() [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return self.greenlet.switch() [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] result = function(*args, **kwargs) [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return func(*args, **kwargs) [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] raise e [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] nwinfo = self.network_api.allocate_for_instance( [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.955293] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] created_port_ids = self._update_ports_for_instance( [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] with excutils.save_and_reraise_exception(): [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self.force_reraise() [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] raise self.value [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] updated_port = self._update_port( [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] _ensure_no_port_binding_failure(port) [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.955735] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] raise exception.PortBindingFailed(port_id=port['id']) [ 663.956101] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] nova.exception.PortBindingFailed: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. [ 663.956101] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] [ 663.956101] env[61995]: INFO nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Terminating instance [ 663.956578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Acquiring lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.250583] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.299393] env[61995]: DEBUG nova.network.neutron [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.354520] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794184, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024328} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.354807] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 664.354979] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Moving file from [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f/c1739132-aff3-4644-9522-fc02cf765bc8 to [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8. {{(pid=61995) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 664.355304] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-2729004a-dad0-4b52-9797-28b9839c4b9c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.362278] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 664.362278] env[61995]: value = "task-794185" [ 664.362278] env[61995]: _type = "Task" [ 664.362278] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.373534] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794185, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.415690] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.452122] env[61995]: DEBUG nova.network.neutron [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.871584] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794185, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.041854} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.874081] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] File moved {{(pid=61995) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 664.874316] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Cleaning up location [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 664.874445] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Deleting the datastore file [datastore1] vmware_temp/0f1f5b7f-a6f3-44c2-a347-7413e88b6c2f {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 664.874873] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7234ab12-49ec-4310-a235-7c01fbe356fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.881705] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 664.881705] env[61995]: value = "task-794186" [ 664.881705] env[61995]: _type = "Task" [ 664.881705] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.892083] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794186, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.919307] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Releasing lock "refresh_cache-adb60754-9217-4b1b-b5cd-98553ff5bbbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.919785] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.919956] env[61995]: DEBUG nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.920218] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.946086] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.956594] env[61995]: DEBUG oslo_concurrency.lockutils [req-9a7b1af7-e434-4fe1-8e98-8b122d770349 req-bc4edba8-b8a5-40b6-aeba-d0bbcd6f1e7c service nova] Releasing lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.957019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Acquired lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.957212] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.105736] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af46f84a-23c7-4fa8-a38f-09b05e3f0703 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.113340] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a35f2a6-0b85-4073-847a-619e3c437dd6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.144026] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcb6b28-9aab-457a-a106-72ea27bca36c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.151326] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c72a05-64af-4ba0-af45-72e6ce68aae5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.164710] env[61995]: DEBUG nova.compute.provider_tree [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.391356] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794186, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022019} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.391631] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 665.392379] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ffb1bfa-90ff-4c21-955b-167f8fdecd4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.397290] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 665.397290] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5285b9ca-96b7-e4d4-90b7-03412b31c4ab" [ 665.397290] env[61995]: _type = "Task" [ 665.397290] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.405114] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5285b9ca-96b7-e4d4-90b7-03412b31c4ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.451191] env[61995]: DEBUG nova.network.neutron [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.477398] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.565365] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.667568] env[61995]: DEBUG nova.scheduler.client.report [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.795836] env[61995]: DEBUG nova.compute.manager [req-b8156a1c-e67b-4eca-b41b-81d59950467b req-6a7cc5c1-cd03-4650-a689-95ad5fbbb885 service nova] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Received event network-vif-deleted-328e9ca5-d5be-4008-a06c-74825b4d0e79 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 665.908109] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5285b9ca-96b7-e4d4-90b7-03412b31c4ab, 'name': SearchDatastore_Task, 'duration_secs': 0.008352} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.908211] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.908398] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 665.908697] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43433be9-5767-4d40-a647-1628e2e5a8e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.915764] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 665.915764] env[61995]: value = "task-794187" [ 665.915764] env[61995]: _type = "Task" [ 665.915764] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.923502] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.954228] env[61995]: INFO nova.compute.manager [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] [instance: adb60754-9217-4b1b-b5cd-98553ff5bbbf] Took 1.03 seconds to deallocate network for instance. [ 666.068411] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Releasing lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.068929] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 666.069281] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 666.069605] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce12d196-4055-465f-9d2a-8af87862f896 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.079785] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e271cf-8581-4c6a-9305-6b3c7479816f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.103954] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 303d1ec5-cc8a-4f25-a36f-aec39e856766 could not be found. [ 666.104254] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 666.104454] env[61995]: INFO nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Took 0.04 seconds to destroy the instance on the hypervisor. [ 666.104710] env[61995]: DEBUG oslo.service.loopingcall [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.105038] env[61995]: DEBUG nova.compute.manager [-] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.105131] env[61995]: DEBUG nova.network.neutron [-] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 666.144067] env[61995]: DEBUG nova.network.neutron [-] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.176435] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.176972] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.180178] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.695s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.426877] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794187, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481286} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.426877] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 666.426877] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 666.427343] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7129c786-81a5-4146-810c-4dabcbb2f728 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.433770] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 666.433770] env[61995]: value = "task-794188" [ 666.433770] env[61995]: _type = "Task" [ 666.433770] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.441351] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794188, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.650031] env[61995]: DEBUG nova.network.neutron [-] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.684517] env[61995]: DEBUG nova.compute.utils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.689184] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.690696] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 666.790194] env[61995]: DEBUG nova.policy [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3a181252f16b495a89199a259cb13ce6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1efe2f470e42496095457ecb89478a0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 666.945437] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794188, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07464} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.946181] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 666.947642] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d55f23c-6861-41c0-9f99-77fed0d255de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.978101] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 666.982246] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2808643c-3383-4562-9ba7-21aed74716b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.999099] env[61995]: INFO nova.scheduler.client.report [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Deleted allocations for instance adb60754-9217-4b1b-b5cd-98553ff5bbbf [ 667.012125] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 667.012125] env[61995]: value = "task-794189" [ 667.012125] env[61995]: _type = "Task" [ 667.012125] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.025191] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794189, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.147074] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10a7a99-ad0c-4fed-a2fe-29d019169190 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.154044] env[61995]: INFO nova.compute.manager [-] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Took 1.05 seconds to deallocate network for instance. [ 667.157205] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947f972c-8310-46a5-9a6f-c45a7da09741 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.161047] env[61995]: DEBUG nova.compute.claims [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.161240] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.189099] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e3d9f4-c293-4ca8-8445-5aab9ed0a658 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.192069] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.202014] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921511cc-ba97-4040-9250-53476a64a520 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.216292] env[61995]: DEBUG nova.compute.provider_tree [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.440357] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Successfully created port: f26530b6-aed8-449d-9719-ce52c3cfadcc {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.511879] env[61995]: DEBUG oslo_concurrency.lockutils [None req-77edd4fb-6828-454a-a9f9-e43501ad6539 tempest-ImagesOneServerNegativeTestJSON-2080042216 tempest-ImagesOneServerNegativeTestJSON-2080042216-project-member] Lock "adb60754-9217-4b1b-b5cd-98553ff5bbbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.722s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.522823] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794189, 'name': ReconfigVM_Task, 'duration_secs': 0.276329} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.524558] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc/2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 667.527092] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abc50b7a-c386-43c0-bebb-496f4c78826c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.533641] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 667.533641] env[61995]: value = "task-794190" [ 667.533641] env[61995]: _type = "Task" [ 667.533641] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.542612] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794190, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.723079] env[61995]: DEBUG nova.scheduler.client.report [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.016757] env[61995]: DEBUG nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.044620] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794190, 'name': Rename_Task, 'duration_secs': 0.128076} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.045072] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 668.045432] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9456f889-db19-4a3a-8d42-11aef4eb93ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.054016] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Waiting for the task: (returnval){ [ 668.054016] env[61995]: value = "task-794191" [ 668.054016] env[61995]: _type = "Task" [ 668.054016] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.060444] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.205220] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.228440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.049s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.229156] env[61995]: ERROR nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Traceback (most recent call last): [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self.driver.spawn(context, instance, image_meta, [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] vm_ref = self.build_virtual_machine(instance, [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.229156] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] for vif in network_info: [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return self._sync_wrapper(fn, *args, **kwargs) [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self.wait() [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self[:] = self._gt.wait() [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return self._exit_event.wait() [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] result = hub.switch() [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.229438] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return self.greenlet.switch() [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] result = function(*args, **kwargs) [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] return func(*args, **kwargs) [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] raise e [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] nwinfo = self.network_api.allocate_for_instance( [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] created_port_ids = self._update_ports_for_instance( [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] with excutils.save_and_reraise_exception(): [ 668.229732] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] self.force_reraise() [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] raise self.value [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] updated_port = self._update_port( [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] _ensure_no_port_binding_failure(port) [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] raise exception.PortBindingFailed(port_id=port['id']) [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] nova.exception.PortBindingFailed: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. [ 668.230026] env[61995]: ERROR nova.compute.manager [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] [ 668.230285] env[61995]: DEBUG nova.compute.utils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.235018] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.235018] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.235018] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.235269] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.235269] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.235269] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.235269] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.235269] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.235390] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.235467] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.235578] env[61995]: DEBUG nova.virt.hardware [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.235919] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.204s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.239403] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63819a4a-3740-4b91-9ba0-aecc3d983355 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.243942] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Build of instance 596463d7-04ca-4f29-8adf-c451310aaf15 was re-scheduled: Binding failed for port 8f75df4e-c4cc-4102-86b9-61563de33e05, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.244468] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.244730] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Acquiring lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.244838] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Acquired lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.244993] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.255387] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ea5155-1031-4b23-823b-19069bc3464d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.542264] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.565017] env[61995]: DEBUG oslo_vmware.api [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Task: {'id': task-794191, 'name': PowerOnVM_Task, 'duration_secs': 0.485353} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.565017] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 668.565017] env[61995]: DEBUG nova.compute.manager [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 668.565017] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c17e0a-7a6b-4cf2-b3e6-91ec7fe01de3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.773806] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.843070] env[61995]: DEBUG nova.compute.manager [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Received event network-changed-f26530b6-aed8-449d-9719-ce52c3cfadcc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 668.843070] env[61995]: DEBUG nova.compute.manager [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Refreshing instance network info cache due to event network-changed-f26530b6-aed8-449d-9719-ce52c3cfadcc. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 668.843070] env[61995]: DEBUG oslo_concurrency.lockutils [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] Acquiring lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.843070] env[61995]: DEBUG oslo_concurrency.lockutils [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] Acquired lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.843070] env[61995]: DEBUG nova.network.neutron [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Refreshing network info cache for port f26530b6-aed8-449d-9719-ce52c3cfadcc {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 668.953048] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.088077] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.124226] env[61995]: ERROR nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. [ 669.124226] env[61995]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.124226] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.124226] env[61995]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.124226] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.124226] env[61995]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.124226] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.124226] env[61995]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.124226] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.124226] env[61995]: ERROR nova.compute.manager self.force_reraise() [ 669.124226] env[61995]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.124226] env[61995]: ERROR nova.compute.manager raise self.value [ 669.124226] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.124226] env[61995]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.124226] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.124226] env[61995]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.126247] env[61995]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.126247] env[61995]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.126247] env[61995]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. [ 669.126247] env[61995]: ERROR nova.compute.manager [ 669.126247] env[61995]: Traceback (most recent call last): [ 669.126247] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.126247] env[61995]: listener.cb(fileno) [ 669.126247] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.126247] env[61995]: result = function(*args, **kwargs) [ 669.126247] env[61995]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.126247] env[61995]: return func(*args, **kwargs) [ 669.126247] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.126247] env[61995]: raise e [ 669.126247] env[61995]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.126247] env[61995]: nwinfo = self.network_api.allocate_for_instance( [ 669.126247] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.126247] env[61995]: created_port_ids = self._update_ports_for_instance( [ 669.126247] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.126247] env[61995]: with excutils.save_and_reraise_exception(): [ 669.126247] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.126247] env[61995]: self.force_reraise() [ 669.126247] env[61995]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.126247] env[61995]: raise self.value [ 669.126247] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.126247] env[61995]: updated_port = self._update_port( [ 669.126247] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.126247] env[61995]: _ensure_no_port_binding_failure(port) [ 669.126247] env[61995]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.126247] env[61995]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.126882] env[61995]: nova.exception.PortBindingFailed: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. [ 669.126882] env[61995]: Removing descriptor: 14 [ 669.126882] env[61995]: ERROR nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Traceback (most recent call last): [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] yield resources [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self.driver.spawn(context, instance, image_meta, [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.126882] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] vm_ref = self.build_virtual_machine(instance, [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] for vif in network_info: [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return self._sync_wrapper(fn, *args, **kwargs) [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self.wait() [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self[:] = self._gt.wait() [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return self._exit_event.wait() [ 669.127167] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] result = hub.switch() [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return self.greenlet.switch() [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] result = function(*args, **kwargs) [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return func(*args, **kwargs) [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] raise e [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] nwinfo = self.network_api.allocate_for_instance( [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.127453] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] created_port_ids = self._update_ports_for_instance( [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] with excutils.save_and_reraise_exception(): [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self.force_reraise() [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] raise self.value [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] updated_port = self._update_port( [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] _ensure_no_port_binding_failure(port) [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.127781] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] raise exception.PortBindingFailed(port_id=port['id']) [ 669.128116] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] nova.exception.PortBindingFailed: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. [ 669.128116] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] [ 669.128116] env[61995]: INFO nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Terminating instance [ 669.130720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.184224] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fd1ee0-12a1-472d-a5f6-80bfa07b86f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.193405] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360554dc-d568-466f-9618-82b406b7abe9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.229661] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af50a0d5-7cf0-44c4-b131-c25d41152225 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.238605] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c95887f-2054-4a38-8f53-aadda72b9f46 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.253468] env[61995]: DEBUG nova.compute.provider_tree [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.383967] env[61995]: DEBUG nova.network.neutron [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.456251] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Releasing lock "refresh_cache-596463d7-04ca-4f29-8adf-c451310aaf15" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.456648] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 669.456992] env[61995]: DEBUG nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.457254] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.466488] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "4da426a1-7af9-4e4e-82ed-c45687a19b86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.466852] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.499831] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.526568] env[61995]: DEBUG nova.network.neutron [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.759783] env[61995]: DEBUG nova.scheduler.client.report [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.935946] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.936306] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.936549] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.936751] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.936935] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.940577] env[61995]: INFO nova.compute.manager [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Terminating instance [ 669.941322] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "refresh_cache-2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.941490] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquired lock "refresh_cache-2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.941662] env[61995]: DEBUG nova.network.neutron [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.004017] env[61995]: DEBUG nova.network.neutron [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.031088] env[61995]: DEBUG oslo_concurrency.lockutils [req-52d76836-7c6e-46f2-9ce3-c28cd3b88b00 req-7e946174-d315-4a5d-8249-f9286f841e20 service nova] Releasing lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.031471] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquired lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.031652] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.267099] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.267099] env[61995]: ERROR nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. [ 670.267099] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Traceback (most recent call last): [ 670.267099] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.267099] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self.driver.spawn(context, instance, image_meta, [ 670.267099] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.267099] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.267099] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.267099] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] vm_ref = self.build_virtual_machine(instance, [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] for vif in network_info: [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return self._sync_wrapper(fn, *args, **kwargs) [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self.wait() [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self[:] = self._gt.wait() [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return self._exit_event.wait() [ 670.267482] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] result = hub.switch() [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return self.greenlet.switch() [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] result = function(*args, **kwargs) [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] return func(*args, **kwargs) [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] raise e [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] nwinfo = self.network_api.allocate_for_instance( [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.267831] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] created_port_ids = self._update_ports_for_instance( [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] with excutils.save_and_reraise_exception(): [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] self.force_reraise() [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] raise self.value [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] updated_port = self._update_port( [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] _ensure_no_port_binding_failure(port) [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.268362] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] raise exception.PortBindingFailed(port_id=port['id']) [ 670.268672] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] nova.exception.PortBindingFailed: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. [ 670.268672] env[61995]: ERROR nova.compute.manager [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] [ 670.268672] env[61995]: DEBUG nova.compute.utils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.269286] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.138s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.269510] env[61995]: DEBUG nova.objects.instance [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 670.272123] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Build of instance 3252888a-31ec-4d93-a8e2-b67aa0bab1e8 was re-scheduled: Binding failed for port 40f80d5f-ab21-4529-8c1a-c8f59f288ea9, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.272525] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.272744] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Acquiring lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.272887] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Acquired lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.273433] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.461440] env[61995]: DEBUG nova.network.neutron [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.508280] env[61995]: INFO nova.compute.manager [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] [instance: 596463d7-04ca-4f29-8adf-c451310aaf15] Took 1.05 seconds to deallocate network for instance. [ 670.518496] env[61995]: DEBUG nova.network.neutron [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.556366] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.618295] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.792051] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.838790] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.867023] env[61995]: DEBUG nova.compute.manager [req-c2136542-ea28-4c87-9307-c6a76c48a652 req-58d68f1e-d902-45fc-b797-b35621411940 service nova] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Received event network-vif-deleted-f26530b6-aed8-449d-9719-ce52c3cfadcc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.023852] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Releasing lock "refresh_cache-2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.024862] env[61995]: DEBUG nova.compute.manager [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.024862] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 671.025523] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43de621-2e56-4e8b-937c-3ed076a3477d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.033762] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 671.034052] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1255635d-c2a7-4605-9819-e1ec183183c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.040247] env[61995]: DEBUG oslo_vmware.api [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 671.040247] env[61995]: value = "task-794192" [ 671.040247] env[61995]: _type = "Task" [ 671.040247] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.048160] env[61995]: DEBUG oslo_vmware.api [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.120828] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Releasing lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.121271] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.121474] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 671.121752] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a72a78f9-9226-4584-a13c-c9b0b26ec7e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.131210] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d331ce9c-96a0-4c4a-b4f1-05a589be4e22 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.153541] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f5024ce-5a48-4d27-a412-35149ee843f3 could not be found. [ 671.153762] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 671.153945] env[61995]: INFO nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 671.154217] env[61995]: DEBUG oslo.service.loopingcall [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.154454] env[61995]: DEBUG nova.compute.manager [-] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.154557] env[61995]: DEBUG nova.network.neutron [-] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.181086] env[61995]: DEBUG nova.network.neutron [-] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.282011] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3149e98a-afb1-45e8-acc2-c4390e1e2ba3 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.283179] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.096s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.340975] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Releasing lock "refresh_cache-3252888a-31ec-4d93-a8e2-b67aa0bab1e8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.341247] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.341409] env[61995]: DEBUG nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.341576] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.360412] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.545179] env[61995]: INFO nova.scheduler.client.report [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Deleted allocations for instance 596463d7-04ca-4f29-8adf-c451310aaf15 [ 671.557036] env[61995]: DEBUG oslo_vmware.api [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794192, 'name': PowerOffVM_Task, 'duration_secs': 0.182848} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.558238] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 671.558383] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 671.558779] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1817322f-1000-4238-8101-bdf7c7918677 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.591423] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 671.591637] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 671.591813] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleting the datastore file [datastore1] 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 671.592111] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a06b4f3b-c61b-4c1a-9bab-cd84faa24000 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.597882] env[61995]: DEBUG oslo_vmware.api [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for the task: (returnval){ [ 671.597882] env[61995]: value = "task-794194" [ 671.597882] env[61995]: _type = "Task" [ 671.597882] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.606055] env[61995]: DEBUG oslo_vmware.api [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.684615] env[61995]: DEBUG nova.network.neutron [-] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.862673] env[61995]: DEBUG nova.network.neutron [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.060282] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88a7a2c7-70ff-435e-a4ea-6082d153479b tempest-ServersWithSpecificFlavorTestJSON-2011492185 tempest-ServersWithSpecificFlavorTestJSON-2011492185-project-member] Lock "596463d7-04ca-4f29-8adf-c451310aaf15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.845s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.112744] env[61995]: DEBUG oslo_vmware.api [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Task: {'id': task-794194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.105068} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.113009] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 672.113204] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 672.113365] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 672.113554] env[61995]: INFO nova.compute.manager [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Took 1.09 seconds to destroy the instance on the hypervisor. [ 672.113761] env[61995]: DEBUG oslo.service.loopingcall [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.113945] env[61995]: DEBUG nova.compute.manager [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.114055] env[61995]: DEBUG nova.network.neutron [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.137842] env[61995]: DEBUG nova.network.neutron [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.187730] env[61995]: INFO nova.compute.manager [-] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Took 1.03 seconds to deallocate network for instance. [ 672.191889] env[61995]: DEBUG nova.compute.claims [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Aborting claim: {{(pid=61995) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.192090] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.220175] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e400beb-9442-4c5c-af8e-b4e290a8e140 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.230873] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ba4665-c3c1-4401-8b05-912bba280da9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.274772] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2dc8fd0-6744-4bbb-aaff-7bd0f4bf0d78 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.282989] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a129d386-e0ca-4bdd-911b-49671cc04f2e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.299148] env[61995]: DEBUG nova.compute.provider_tree [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.365303] env[61995]: INFO nova.compute.manager [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] [instance: 3252888a-31ec-4d93-a8e2-b67aa0bab1e8] Took 1.02 seconds to deallocate network for instance. [ 672.566624] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 672.640408] env[61995]: DEBUG nova.network.neutron [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.801687] env[61995]: DEBUG nova.scheduler.client.report [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.085390] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.143180] env[61995]: INFO nova.compute.manager [-] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Took 1.03 seconds to deallocate network for instance. [ 673.307663] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.024s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.308317] env[61995]: ERROR nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Traceback (most recent call last): [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self.driver.spawn(context, instance, image_meta, [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] vm_ref = self.build_virtual_machine(instance, [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.308317] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] for vif in network_info: [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return self._sync_wrapper(fn, *args, **kwargs) [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self.wait() [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self[:] = self._gt.wait() [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return self._exit_event.wait() [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] result = hub.switch() [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 673.308638] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return self.greenlet.switch() [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] result = function(*args, **kwargs) [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] return func(*args, **kwargs) [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] raise e [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] nwinfo = self.network_api.allocate_for_instance( [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] created_port_ids = self._update_ports_for_instance( [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] with excutils.save_and_reraise_exception(): [ 673.309055] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] self.force_reraise() [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] raise self.value [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] updated_port = self._update_port( [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] _ensure_no_port_binding_failure(port) [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] raise exception.PortBindingFailed(port_id=port['id']) [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] nova.exception.PortBindingFailed: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. [ 673.309468] env[61995]: ERROR nova.compute.manager [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] [ 673.309814] env[61995]: DEBUG nova.compute.utils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.310270] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.575s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.313804] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Build of instance 862694a5-63f8-4e2d-982a-329f4a071d29 was re-scheduled: Binding failed for port 020351b2-23a4-472f-96fd-d7ecf48067f0, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.314855] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.315652] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.315878] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquired lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.316075] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.398825] env[61995]: INFO nova.scheduler.client.report [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Deleted allocations for instance 3252888a-31ec-4d93-a8e2-b67aa0bab1e8 [ 673.650318] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.847369] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.910602] env[61995]: DEBUG oslo_concurrency.lockutils [None req-62b1c260-08c7-4ed1-a880-f716d88a0a80 tempest-ServerExternalEventsTest-1852514373 tempest-ServerExternalEventsTest-1852514373-project-member] Lock "3252888a-31ec-4d93-a8e2-b67aa0bab1e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.533s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.983367] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.192023] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c0302a-10bd-41d1-b6dd-3db6fbfa9f5d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.201073] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f144ad6-2253-4968-8b17-56e0ec338af0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.236422] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bbe982-06bc-4507-9707-9100e3126151 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.243777] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451474da-445c-48af-a3d8-1aa97ea61bc8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.256742] env[61995]: DEBUG nova.compute.provider_tree [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.416313] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 674.486468] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Releasing lock "refresh_cache-862694a5-63f8-4e2d-982a-329f4a071d29" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.486812] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.486998] env[61995]: DEBUG nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.487191] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.509349] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.760058] env[61995]: DEBUG nova.scheduler.client.report [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.954378] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.011886] env[61995]: DEBUG nova.network.neutron [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.266803] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.267467] env[61995]: ERROR nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Traceback (most recent call last): [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self.driver.spawn(context, instance, image_meta, [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] vm_ref = self.build_virtual_machine(instance, [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.267467] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] for vif in network_info: [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return self._sync_wrapper(fn, *args, **kwargs) [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self.wait() [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self[:] = self._gt.wait() [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return self._exit_event.wait() [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] result = hub.switch() [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.267752] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return self.greenlet.switch() [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] result = function(*args, **kwargs) [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] return func(*args, **kwargs) [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] raise e [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] nwinfo = self.network_api.allocate_for_instance( [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] created_port_ids = self._update_ports_for_instance( [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] with excutils.save_and_reraise_exception(): [ 675.268064] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] self.force_reraise() [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] raise self.value [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] updated_port = self._update_port( [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] _ensure_no_port_binding_failure(port) [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] raise exception.PortBindingFailed(port_id=port['id']) [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] nova.exception.PortBindingFailed: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. [ 675.268346] env[61995]: ERROR nova.compute.manager [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] [ 675.268633] env[61995]: DEBUG nova.compute.utils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.269538] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.644s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.273106] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Build of instance 7bdf0b12-25d2-45a5-af57-b1bb31148f3d was re-scheduled: Binding failed for port f761173d-e9c8-4514-9ee8-2ad47f90a6b3, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.273543] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.273726] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Acquiring lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.273889] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Acquired lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.274028] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.517441] env[61995]: INFO nova.compute.manager [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: 862694a5-63f8-4e2d-982a-329f4a071d29] Took 1.03 seconds to deallocate network for instance. [ 675.810102] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.972151] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.209344] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb2899a-6dd8-4a26-bf9b-5f99e039de9a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.218497] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fbf156e-78bc-4c7e-bc4b-3004f0ec9b68 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.840252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Releasing lock "refresh_cache-7bdf0b12-25d2-45a5-af57-b1bb31148f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.840630] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.840768] env[61995]: DEBUG nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.840838] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.845809] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cf5fe7-5f1e-4279-aaef-15133a7bf92c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.853357] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742bc785-2894-4694-b2ca-d93f06cb0e48 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.866714] env[61995]: DEBUG nova.compute.provider_tree [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.871835] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.877044] env[61995]: INFO nova.scheduler.client.report [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Deleted allocations for instance 862694a5-63f8-4e2d-982a-329f4a071d29 [ 677.370110] env[61995]: DEBUG nova.scheduler.client.report [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.374070] env[61995]: DEBUG nova.network.neutron [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.384455] env[61995]: DEBUG oslo_concurrency.lockutils [None req-969bfe68-820c-4733-9e40-adef21e2476d tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "862694a5-63f8-4e2d-982a-329f4a071d29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.166s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.883228] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.611s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.883228] env[61995]: ERROR nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. [ 677.883228] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Traceback (most recent call last): [ 677.883228] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.883228] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self.driver.spawn(context, instance, image_meta, [ 677.883228] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.883228] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.883228] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.883228] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] vm_ref = self.build_virtual_machine(instance, [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] for vif in network_info: [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return self._sync_wrapper(fn, *args, **kwargs) [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self.wait() [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self[:] = self._gt.wait() [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return self._exit_event.wait() [ 677.883638] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] result = hub.switch() [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return self.greenlet.switch() [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] result = function(*args, **kwargs) [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] return func(*args, **kwargs) [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] raise e [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] nwinfo = self.network_api.allocate_for_instance( [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.883938] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] created_port_ids = self._update_ports_for_instance( [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] with excutils.save_and_reraise_exception(): [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] self.force_reraise() [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] raise self.value [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] updated_port = self._update_port( [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] _ensure_no_port_binding_failure(port) [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.884326] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] raise exception.PortBindingFailed(port_id=port['id']) [ 677.884606] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] nova.exception.PortBindingFailed: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. [ 677.884606] env[61995]: ERROR nova.compute.manager [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] [ 677.884606] env[61995]: DEBUG nova.compute.utils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.885985] env[61995]: INFO nova.compute.manager [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] [instance: 7bdf0b12-25d2-45a5-af57-b1bb31148f3d] Took 1.04 seconds to deallocate network for instance. [ 677.889060] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Build of instance 5ca64ce9-9856-4e65-a2e7-af11ae0211da was re-scheduled: Binding failed for port d861c7f2-f048-4f98-b1a3-61ac952502d2, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.889829] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.890175] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Acquiring lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.890465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Acquired lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.890712] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.891817] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.949s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.894306] env[61995]: INFO nova.compute.claims [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.897108] env[61995]: DEBUG nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 678.433913] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.437909] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.552085] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.940465] env[61995]: INFO nova.scheduler.client.report [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Deleted allocations for instance 7bdf0b12-25d2-45a5-af57-b1bb31148f3d [ 679.055592] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Releasing lock "refresh_cache-5ca64ce9-9856-4e65-a2e7-af11ae0211da" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.056447] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 679.056447] env[61995]: DEBUG nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.056447] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.084279] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.327317] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bab258f-945e-44e6-8684-6734dd6cf446 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.335607] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abdc108-68be-4088-834b-b38c10008b9d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.368719] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f94cf9b-ad8b-404e-94cf-5f5baff3616a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.376419] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5930f229-91c7-4d7b-ae58-be5d509c7383 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.390458] env[61995]: DEBUG nova.compute.provider_tree [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.459466] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e5dcfde-ef7e-4a5c-bb4e-d5f872d4844c tempest-ImagesNegativeTestJSON-1932799863 tempest-ImagesNegativeTestJSON-1932799863-project-member] Lock "7bdf0b12-25d2-45a5-af57-b1bb31148f3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.200s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.588453] env[61995]: DEBUG nova.network.neutron [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.894420] env[61995]: DEBUG nova.scheduler.client.report [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.967980] env[61995]: DEBUG nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.091909] env[61995]: INFO nova.compute.manager [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] [instance: 5ca64ce9-9856-4e65-a2e7-af11ae0211da] Took 1.04 seconds to deallocate network for instance. [ 680.404099] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.404099] env[61995]: DEBUG nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 680.406990] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.537s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.409830] env[61995]: INFO nova.compute.claims [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.493433] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.915667] env[61995]: DEBUG nova.compute.utils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.925670] env[61995]: DEBUG nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.925670] env[61995]: DEBUG nova.network.neutron [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 681.030685] env[61995]: DEBUG nova.policy [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '232abb0cac4b49ce8a6ca2647701aad8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b368dcb48f8e4b18bee5814674080022', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 681.143065] env[61995]: INFO nova.scheduler.client.report [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Deleted allocations for instance 5ca64ce9-9856-4e65-a2e7-af11ae0211da [ 681.371088] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.371088] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.420155] env[61995]: DEBUG nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 681.478426] env[61995]: DEBUG nova.network.neutron [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Successfully created port: c0c4eccc-870f-40e6-bbe0-f8c11172e6df {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.664838] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a4e83a7f-a51d-4341-8904-ff504ba957fe tempest-ServersV294TestFqdnHostnames-30897634 tempest-ServersV294TestFqdnHostnames-30897634-project-member] Lock "5ca64ce9-9856-4e65-a2e7-af11ae0211da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.867s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.860533] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca92dd3-7faa-4bb0-9837-b110d9f36eec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.869459] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0c21ce-00db-4e4b-be0c-63c8e6a9c4d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.904936] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c19d07-7c05-4e61-b0dd-253a4be27448 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.912796] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8308118-93ca-4f5c-af90-9ea832890081 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.935533] env[61995]: DEBUG nova.compute.provider_tree [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.166026] env[61995]: DEBUG nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.396379] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.396680] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.438902] env[61995]: DEBUG nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 682.441617] env[61995]: DEBUG nova.scheduler.client.report [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.482270] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.482270] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.482270] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.482419] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.482419] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.482419] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.482419] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.482419] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.482544] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.482544] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.482544] env[61995]: DEBUG nova.virt.hardware [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.483112] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da020eb0-a09b-43a1-ad40-53e1bd034ec2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.495872] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b79ab3-53f2-4651-bece-213881df5520 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.692398] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.949151] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.949151] env[61995]: DEBUG nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.951238] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.790s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.412033] env[61995]: DEBUG nova.network.neutron [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Successfully updated port: c0c4eccc-870f-40e6-bbe0-f8c11172e6df {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 683.450859] env[61995]: DEBUG nova.compute.manager [req-cc040476-875f-4f65-8670-5119fbaf683e req-c623eeff-e9b8-44df-855e-98dfb492f76a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Received event network-vif-plugged-c0c4eccc-870f-40e6-bbe0-f8c11172e6df {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 683.450859] env[61995]: DEBUG oslo_concurrency.lockutils [req-cc040476-875f-4f65-8670-5119fbaf683e req-c623eeff-e9b8-44df-855e-98dfb492f76a service nova] Acquiring lock "06496d6a-324d-41d4-b3e5-52e3d93636bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.450859] env[61995]: DEBUG oslo_concurrency.lockutils [req-cc040476-875f-4f65-8670-5119fbaf683e req-c623eeff-e9b8-44df-855e-98dfb492f76a service nova] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.450859] env[61995]: DEBUG oslo_concurrency.lockutils [req-cc040476-875f-4f65-8670-5119fbaf683e req-c623eeff-e9b8-44df-855e-98dfb492f76a service nova] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.450859] env[61995]: DEBUG nova.compute.manager [req-cc040476-875f-4f65-8670-5119fbaf683e req-c623eeff-e9b8-44df-855e-98dfb492f76a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] No waiting events found dispatching network-vif-plugged-c0c4eccc-870f-40e6-bbe0-f8c11172e6df {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 683.451018] env[61995]: WARNING nova.compute.manager [req-cc040476-875f-4f65-8670-5119fbaf683e req-c623eeff-e9b8-44df-855e-98dfb492f76a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Received unexpected event network-vif-plugged-c0c4eccc-870f-40e6-bbe0-f8c11172e6df for instance with vm_state building and task_state spawning. [ 683.456568] env[61995]: DEBUG nova.compute.utils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 683.461439] env[61995]: DEBUG nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 683.462450] env[61995]: DEBUG nova.network.neutron [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 683.529317] env[61995]: DEBUG nova.policy [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3a181252f16b495a89199a259cb13ce6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1efe2f470e42496095457ecb89478a0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 683.869245] env[61995]: DEBUG nova.network.neutron [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Successfully created port: b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.873821] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b69d90-0dee-4a57-95d8-61d5b0eab87a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.882230] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756cf69f-fcb0-43b2-88d8-06aa2bceb820 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.921858] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "refresh_cache-06496d6a-324d-41d4-b3e5-52e3d93636bb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.921858] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquired lock "refresh_cache-06496d6a-324d-41d4-b3e5-52e3d93636bb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.922628] env[61995]: DEBUG nova.network.neutron [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 683.924735] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236f7b43-321b-4488-ae6f-0837ae8303b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.933834] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d5933e-9390-491c-a0f3-eaf11570b8c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.951488] env[61995]: DEBUG nova.compute.provider_tree [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.960853] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 683.961087] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 683.965885] env[61995]: DEBUG nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 684.454792] env[61995]: DEBUG nova.scheduler.client.report [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.468594] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.470215] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 684.470215] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 684.503109] env[61995]: DEBUG nova.network.neutron [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.892650] env[61995]: DEBUG nova.network.neutron [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Updating instance_info_cache with network_info: [{"id": "c0c4eccc-870f-40e6-bbe0-f8c11172e6df", "address": "fa:16:3e:49:7a:ec", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0c4eccc-87", "ovs_interfaceid": "c0c4eccc-870f-40e6-bbe0-f8c11172e6df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.960191] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.960888] env[61995]: ERROR nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Traceback (most recent call last): [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self.driver.spawn(context, instance, image_meta, [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] vm_ref = self.build_virtual_machine(instance, [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.960888] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] for vif in network_info: [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return self._sync_wrapper(fn, *args, **kwargs) [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self.wait() [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self[:] = self._gt.wait() [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return self._exit_event.wait() [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] result = hub.switch() [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.961206] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return self.greenlet.switch() [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] result = function(*args, **kwargs) [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] return func(*args, **kwargs) [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] raise e [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] nwinfo = self.network_api.allocate_for_instance( [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] created_port_ids = self._update_ports_for_instance( [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] with excutils.save_and_reraise_exception(): [ 684.961483] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] self.force_reraise() [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] raise self.value [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] updated_port = self._update_port( [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] _ensure_no_port_binding_failure(port) [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] raise exception.PortBindingFailed(port_id=port['id']) [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] nova.exception.PortBindingFailed: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. [ 684.961799] env[61995]: ERROR nova.compute.manager [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] [ 684.962049] env[61995]: DEBUG nova.compute.utils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.963938] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Build of instance 303d1ec5-cc8a-4f25-a36f-aec39e856766 was re-scheduled: Binding failed for port 328e9ca5-d5be-4008-a06c-74825b4d0e79, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.965407] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.965407] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Acquiring lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.965681] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Acquired lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.965965] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.967974] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.426s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.969627] env[61995]: INFO nova.compute.claims [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.981619] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 684.981924] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 684.982072] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 684.987321] env[61995]: DEBUG nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 685.015160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.015160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.015160] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 685.015160] env[61995]: DEBUG nova.objects.instance [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lazy-loading 'info_cache' on Instance uuid 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 685.024807] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 685.025052] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 685.025519] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.026956] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 685.027286] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.029035] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 685.029035] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 685.029035] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 685.029035] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 685.029035] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 685.029323] env[61995]: DEBUG nova.virt.hardware [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 685.029846] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3e97a9-6b5c-442a-9c69-e5629f3766d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.040709] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a6008a-9171-45b2-ae78-e0d2e60ce235 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.395898] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Releasing lock "refresh_cache-06496d6a-324d-41d4-b3e5-52e3d93636bb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.396268] env[61995]: DEBUG nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Instance network_info: |[{"id": "c0c4eccc-870f-40e6-bbe0-f8c11172e6df", "address": "fa:16:3e:49:7a:ec", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0c4eccc-87", "ovs_interfaceid": "c0c4eccc-870f-40e6-bbe0-f8c11172e6df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 685.396759] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:7a:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0c4eccc-870f-40e6-bbe0-f8c11172e6df', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 685.409980] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Creating folder: Project (b368dcb48f8e4b18bee5814674080022). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 685.410629] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcee1f50-b9b5-4e17-9b1c-611f4a07f25c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.421972] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Created folder: Project (b368dcb48f8e4b18bee5814674080022) in parent group-v185203. [ 685.422186] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Creating folder: Instances. Parent ref: group-v185209. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 685.422492] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4349e08-90fc-4807-a538-477b4a114151 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.431295] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Created folder: Instances in parent group-v185209. [ 685.431454] env[61995]: DEBUG oslo.service.loopingcall [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 685.431595] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 685.431788] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a81ebe6a-7ee9-4ae7-904f-45ce4643b377 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.453706] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 685.453706] env[61995]: value = "task-794197" [ 685.453706] env[61995]: _type = "Task" [ 685.453706] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.461328] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794197, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.507551] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.527088] env[61995]: DEBUG nova.compute.manager [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Received event network-changed-c0c4eccc-870f-40e6-bbe0-f8c11172e6df {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 685.527088] env[61995]: DEBUG nova.compute.manager [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Refreshing instance network info cache due to event network-changed-c0c4eccc-870f-40e6-bbe0-f8c11172e6df. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 685.527088] env[61995]: DEBUG oslo_concurrency.lockutils [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] Acquiring lock "refresh_cache-06496d6a-324d-41d4-b3e5-52e3d93636bb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.527088] env[61995]: DEBUG oslo_concurrency.lockutils [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] Acquired lock "refresh_cache-06496d6a-324d-41d4-b3e5-52e3d93636bb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.527088] env[61995]: DEBUG nova.network.neutron [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Refreshing network info cache for port c0c4eccc-870f-40e6-bbe0-f8c11172e6df {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 685.735092] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.965868] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794197, 'name': CreateVM_Task, 'duration_secs': 0.349397} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.966797] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 685.987029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.987029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.987029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 685.987029] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6e4025c-214a-4296-abad-3bee09be66ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.991448] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 685.991448] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52adae5d-f808-62dc-8b24-efe235a26665" [ 685.991448] env[61995]: _type = "Task" [ 685.991448] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.002194] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52adae5d-f808-62dc-8b24-efe235a26665, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.047077] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.239536] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Releasing lock "refresh_cache-303d1ec5-cc8a-4f25-a36f-aec39e856766" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.239751] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 686.240187] env[61995]: DEBUG nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.240403] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 686.267475] env[61995]: DEBUG nova.network.neutron [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Successfully updated port: b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 686.269510] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.373167] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba88828-3feb-48be-98fc-6b8231745bba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.380601] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0284e4-3f61-4d37-a7c0-4849db9b0f78 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.411463] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1378b4c-599d-4be7-acab-cc11c618ed8e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.419871] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa8d0b8-4203-4e5c-b9f4-64c3bdc5f64f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.424961] env[61995]: DEBUG nova.network.neutron [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Updated VIF entry in instance network info cache for port c0c4eccc-870f-40e6-bbe0-f8c11172e6df. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 686.425353] env[61995]: DEBUG nova.network.neutron [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Updating instance_info_cache with network_info: [{"id": "c0c4eccc-870f-40e6-bbe0-f8c11172e6df", "address": "fa:16:3e:49:7a:ec", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0c4eccc-87", "ovs_interfaceid": "c0c4eccc-870f-40e6-bbe0-f8c11172e6df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.434582] env[61995]: DEBUG nova.compute.provider_tree [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.502869] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52adae5d-f808-62dc-8b24-efe235a26665, 'name': SearchDatastore_Task, 'duration_secs': 0.010601} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.503183] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.503412] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 686.503636] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.503780] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.503958] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 686.504227] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23de9a54-d219-40b6-ae21-6273940dd766 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.515967] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 686.516212] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 686.517232] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caba3db6-de48-4382-9494-f1826a923b36 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.522215] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 686.522215] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]528c84fc-953a-3b4b-3bba-8f1e48961222" [ 686.522215] env[61995]: _type = "Task" [ 686.522215] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.529345] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528c84fc-953a-3b4b-3bba-8f1e48961222, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.616697] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.774742] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "refresh_cache-d5a54c46-1a45-4bd5-8694-0defc757c455" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.774742] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquired lock "refresh_cache-d5a54c46-1a45-4bd5-8694-0defc757c455" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.774742] env[61995]: DEBUG nova.network.neutron [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 686.774742] env[61995]: DEBUG nova.network.neutron [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.940015] env[61995]: DEBUG oslo_concurrency.lockutils [req-3e366e01-2989-49a7-b3cc-72c39fef9e58 req-6c49ad5b-e9e7-4504-af1c-889f07cccf9a service nova] Releasing lock "refresh_cache-06496d6a-324d-41d4-b3e5-52e3d93636bb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.940015] env[61995]: DEBUG nova.scheduler.client.report [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.036784] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528c84fc-953a-3b4b-3bba-8f1e48961222, 'name': SearchDatastore_Task, 'duration_secs': 0.007707} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.036784] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96a661ed-a27b-46f1-bdd5-dbd8717f7712 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.041540] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 687.041540] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52672321-d8d7-b9de-bca8-5506dd6eb8c5" [ 687.041540] env[61995]: _type = "Task" [ 687.041540] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.049393] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52672321-d8d7-b9de-bca8-5506dd6eb8c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.118701] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.118932] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 687.119186] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 687.119354] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 687.119498] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 687.119642] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 687.119778] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 687.119918] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 687.120056] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 687.120208] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 687.280163] env[61995]: INFO nova.compute.manager [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] [instance: 303d1ec5-cc8a-4f25-a36f-aec39e856766] Took 1.04 seconds to deallocate network for instance. [ 687.334796] env[61995]: DEBUG nova.network.neutron [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.444219] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.444815] env[61995]: DEBUG nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.450127] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.361s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.450338] env[61995]: DEBUG nova.objects.instance [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 687.552841] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52672321-d8d7-b9de-bca8-5506dd6eb8c5, 'name': SearchDatastore_Task, 'duration_secs': 0.008657} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.553302] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.553601] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 06496d6a-324d-41d4-b3e5-52e3d93636bb/06496d6a-324d-41d4-b3e5-52e3d93636bb.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 687.553858] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58228565-4819-4c8b-9901-414a58e783c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.563382] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 687.563382] env[61995]: value = "task-794198" [ 687.563382] env[61995]: _type = "Task" [ 687.563382] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.573914] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794198, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.624083] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.645575] env[61995]: DEBUG nova.network.neutron [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Updating instance_info_cache with network_info: [{"id": "b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0", "address": "fa:16:3e:08:cf:01", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1c0d4ae-53", "ovs_interfaceid": "b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.880140] env[61995]: DEBUG nova.compute.manager [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Received event network-vif-plugged-b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 687.880395] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] Acquiring lock "d5a54c46-1a45-4bd5-8694-0defc757c455-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.880611] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.880783] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.880951] env[61995]: DEBUG nova.compute.manager [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] No waiting events found dispatching network-vif-plugged-b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 687.881136] env[61995]: WARNING nova.compute.manager [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Received unexpected event network-vif-plugged-b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 for instance with vm_state building and task_state spawning. [ 687.881303] env[61995]: DEBUG nova.compute.manager [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Received event network-changed-b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 687.881438] env[61995]: DEBUG nova.compute.manager [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Refreshing instance network info cache due to event network-changed-b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 687.881600] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] Acquiring lock "refresh_cache-d5a54c46-1a45-4bd5-8694-0defc757c455" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.957894] env[61995]: DEBUG nova.compute.utils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.963677] env[61995]: DEBUG nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Not allocating networking since 'none' was specified. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 688.075476] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794198, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460134} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.075949] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 06496d6a-324d-41d4-b3e5-52e3d93636bb/06496d6a-324d-41d4-b3e5-52e3d93636bb.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 688.076970] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 688.076970] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-504ee8fd-e075-4aa6-87de-5321d085b16b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.083301] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 688.083301] env[61995]: value = "task-794199" [ 688.083301] env[61995]: _type = "Task" [ 688.083301] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.092306] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794199, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.152065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Releasing lock "refresh_cache-d5a54c46-1a45-4bd5-8694-0defc757c455" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.153917] env[61995]: DEBUG nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Instance network_info: |[{"id": "b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0", "address": "fa:16:3e:08:cf:01", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1c0d4ae-53", "ovs_interfaceid": "b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 688.153917] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] Acquired lock "refresh_cache-d5a54c46-1a45-4bd5-8694-0defc757c455" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.154094] env[61995]: DEBUG nova.network.neutron [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Refreshing network info cache for port b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 688.156009] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:cf:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 688.165804] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Creating folder: Project (1efe2f470e42496095457ecb89478a0e). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 688.167765] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa26d20f-7e94-4689-90f8-3f9549aa464a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.180139] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Created folder: Project (1efe2f470e42496095457ecb89478a0e) in parent group-v185203. [ 688.180139] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Creating folder: Instances. Parent ref: group-v185212. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 688.180139] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6903b641-776e-4a0b-a3dc-14f4b137ac3a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.191458] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Created folder: Instances in parent group-v185212. [ 688.194026] env[61995]: DEBUG oslo.service.loopingcall [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.194026] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 688.194026] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99287328-56fe-4bed-8336-b1ca33565a95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.214630] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 688.214630] env[61995]: value = "task-794202" [ 688.214630] env[61995]: _type = "Task" [ 688.214630] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.223195] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794202, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.320538] env[61995]: INFO nova.scheduler.client.report [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Deleted allocations for instance 303d1ec5-cc8a-4f25-a36f-aec39e856766 [ 688.463745] env[61995]: DEBUG nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.470793] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4c46762-40fa-4ea7-b634-22789a5d0571 tempest-ServersAdmin275Test-2130864279 tempest-ServersAdmin275Test-2130864279-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.471818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.280s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.595021] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794199, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066138} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.595392] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 688.596276] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b9ec32-271e-43ef-8903-b0e751fa88a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.626588] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] 06496d6a-324d-41d4-b3e5-52e3d93636bb/06496d6a-324d-41d4-b3e5-52e3d93636bb.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 688.627156] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1aa29ccf-f86d-4056-a8c8-13ec23246946 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.651186] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 688.651186] env[61995]: value = "task-794203" [ 688.651186] env[61995]: _type = "Task" [ 688.651186] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.663792] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794203, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.726941] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794202, 'name': CreateVM_Task, 'duration_secs': 0.364536} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.729636] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 688.730158] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.730684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.730684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 688.731259] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55deacb4-1d0a-4e2d-86c0-5aca65fc327f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.736192] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 688.736192] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52981abb-1014-bae5-b0d3-25b1fc4bc0a4" [ 688.736192] env[61995]: _type = "Task" [ 688.736192] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.744078] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52981abb-1014-bae5-b0d3-25b1fc4bc0a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.832282] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee35fa5d-756f-4fbb-92b7-fef639f60c3a tempest-ServersTestJSON-1077751889 tempest-ServersTestJSON-1077751889-project-member] Lock "303d1ec5-cc8a-4f25-a36f-aec39e856766" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.265s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.123960] env[61995]: DEBUG nova.network.neutron [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Updated VIF entry in instance network info cache for port b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 689.124352] env[61995]: DEBUG nova.network.neutron [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Updating instance_info_cache with network_info: [{"id": "b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0", "address": "fa:16:3e:08:cf:01", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1c0d4ae-53", "ovs_interfaceid": "b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.169778] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794203, 'name': ReconfigVM_Task, 'duration_secs': 0.318773} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.173590] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Reconfigured VM instance instance-00000016 to attach disk [datastore1] 06496d6a-324d-41d4-b3e5-52e3d93636bb/06496d6a-324d-41d4-b3e5-52e3d93636bb.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 689.174285] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4d1c9c6-520d-4736-8d7a-f66dac15ab3e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.181339] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 689.181339] env[61995]: value = "task-794204" [ 689.181339] env[61995]: _type = "Task" [ 689.181339] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.192048] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794204, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.250138] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52981abb-1014-bae5-b0d3-25b1fc4bc0a4, 'name': SearchDatastore_Task, 'duration_secs': 0.008801} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.250483] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.251301] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 689.251617] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.252640] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.252640] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 689.252640] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da9a6691-64a0-4d11-ab27-2dbc560ca9d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.264786] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 689.265218] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 689.266446] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5a83867-2438-4072-9e89-19ecd2745371 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.276345] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 689.276345] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52097ea7-a8e9-180e-72e2-a01af0775ab6" [ 689.276345] env[61995]: _type = "Task" [ 689.276345] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.291907] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52097ea7-a8e9-180e-72e2-a01af0775ab6, 'name': SearchDatastore_Task, 'duration_secs': 0.009283} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.292722] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79734c45-766a-4612-ba3d-7c73f9d18140 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.298166] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 689.298166] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e3e794-f40b-4f39-05e0-8d7bc9767c44" [ 689.298166] env[61995]: _type = "Task" [ 689.298166] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.309058] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e3e794-f40b-4f39-05e0-8d7bc9767c44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.338798] env[61995]: DEBUG nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.405106] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1c4030-8cb6-4a9d-b60f-31ab03ea2299 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.413452] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2a7fec-2d67-4092-8d0c-259d3ed96bb9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.444410] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea98809f-590f-4255-a8d6-687663b704b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.455019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3027669e-7f5b-4db7-bfdb-3502e3bc4da5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.467963] env[61995]: DEBUG nova.compute.provider_tree [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.474184] env[61995]: DEBUG nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.506758] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.507039] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.507335] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.507393] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.507518] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.507664] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.507869] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.508093] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.508236] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.508403] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.508572] env[61995]: DEBUG nova.virt.hardware [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.510036] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af35d715-addd-4fb8-bbf0-7abe8dd9727b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.519157] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82273ff-50e7-471b-95b4-3043414578aa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.534913] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.540950] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Creating folder: Project (ff1072ee99634ad28e3c01b403a2b06a). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 689.541296] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd0f37fe-c585-4fe7-9d64-c9c1b2647e20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.554872] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Created folder: Project (ff1072ee99634ad28e3c01b403a2b06a) in parent group-v185203. [ 689.554980] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Creating folder: Instances. Parent ref: group-v185215. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 689.555147] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb5b8b1b-353b-49aa-80fa-c32b7454a3be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.569942] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Created folder: Instances in parent group-v185215. [ 689.570352] env[61995]: DEBUG oslo.service.loopingcall [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.570556] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 689.570765] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82737e8d-1024-4f03-92b4-5c202cf81f64 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.588464] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.588464] env[61995]: value = "task-794207" [ 689.588464] env[61995]: _type = "Task" [ 689.588464] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.596371] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794207, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.627695] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd31a1cd-dcd7-4a4a-bedc-e982012dff95 req-6d45c8ef-7655-4127-a668-6e73f68cb1d6 service nova] Releasing lock "refresh_cache-d5a54c46-1a45-4bd5-8694-0defc757c455" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.691169] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794204, 'name': Rename_Task, 'duration_secs': 0.225005} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.691420] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 689.691657] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c945dd73-31d8-4429-88c1-268b8b9e44f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.698417] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 689.698417] env[61995]: value = "task-794208" [ 689.698417] env[61995]: _type = "Task" [ 689.698417] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.709428] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.809519] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e3e794-f40b-4f39-05e0-8d7bc9767c44, 'name': SearchDatastore_Task, 'duration_secs': 0.009079} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.810427] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.810427] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] d5a54c46-1a45-4bd5-8694-0defc757c455/d5a54c46-1a45-4bd5-8694-0defc757c455.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 689.810427] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c27d5b46-9ab7-4522-b4b6-1e62bf04b51b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.818454] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 689.818454] env[61995]: value = "task-794209" [ 689.818454] env[61995]: _type = "Task" [ 689.818454] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.830023] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794209, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.864108] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.976377] env[61995]: DEBUG nova.scheduler.client.report [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.102958] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794207, 'name': CreateVM_Task, 'duration_secs': 0.301875} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.102958] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 690.103183] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.103609] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.104126] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 690.104443] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbed3f28-a178-43ec-b042-249602a26812 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.111373] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 690.111373] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c10c77-47f4-e650-47ad-829d3ade7402" [ 690.111373] env[61995]: _type = "Task" [ 690.111373] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.120425] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c10c77-47f4-e650-47ad-829d3ade7402, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.209221] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794208, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.331665] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794209, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493224} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.332043] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] d5a54c46-1a45-4bd5-8694-0defc757c455/d5a54c46-1a45-4bd5-8694-0defc757c455.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 690.332274] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.332531] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe4b2d6a-48c3-4e91-ac75-2a65492cd016 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.339599] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 690.339599] env[61995]: value = "task-794210" [ 690.339599] env[61995]: _type = "Task" [ 690.339599] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.353526] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794210, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.483547] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.484272] env[61995]: ERROR nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Traceback (most recent call last): [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self.driver.spawn(context, instance, image_meta, [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] vm_ref = self.build_virtual_machine(instance, [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.484272] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] for vif in network_info: [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return self._sync_wrapper(fn, *args, **kwargs) [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self.wait() [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self[:] = self._gt.wait() [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return self._exit_event.wait() [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] result = hub.switch() [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.484598] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return self.greenlet.switch() [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] result = function(*args, **kwargs) [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] return func(*args, **kwargs) [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] raise e [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] nwinfo = self.network_api.allocate_for_instance( [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] created_port_ids = self._update_ports_for_instance( [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] with excutils.save_and_reraise_exception(): [ 690.486123] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] self.force_reraise() [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] raise self.value [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] updated_port = self._update_port( [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] _ensure_no_port_binding_failure(port) [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] raise exception.PortBindingFailed(port_id=port['id']) [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] nova.exception.PortBindingFailed: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. [ 690.486416] env[61995]: ERROR nova.compute.manager [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] [ 690.486682] env[61995]: DEBUG nova.compute.utils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.486682] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.401s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.488025] env[61995]: INFO nova.compute.claims [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.491392] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Build of instance 9f5024ce-5a48-4d27-a412-35149ee843f3 was re-scheduled: Binding failed for port f26530b6-aed8-449d-9719-ce52c3cfadcc, please check neutron logs for more information. {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.491844] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Unplugging VIFs for instance {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.492262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.492416] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquired lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.492575] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.628294] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c10c77-47f4-e650-47ad-829d3ade7402, 'name': SearchDatastore_Task, 'duration_secs': 0.048966} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.628645] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.629074] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 690.629208] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.629386] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.629552] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.629796] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc83f18d-de95-4b26-96a0-c33b86995db1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.639743] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.639743] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 690.640086] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-374b3495-5a37-4532-8472-c21ba2c747d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.646716] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 690.646716] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52962769-5ea3-688f-21e9-01e1d8305ac5" [ 690.646716] env[61995]: _type = "Task" [ 690.646716] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.657329] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52962769-5ea3-688f-21e9-01e1d8305ac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.709433] env[61995]: DEBUG oslo_vmware.api [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794208, 'name': PowerOnVM_Task, 'duration_secs': 0.56528} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.709718] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 690.709919] env[61995]: INFO nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Took 8.27 seconds to spawn the instance on the hypervisor. [ 690.710108] env[61995]: DEBUG nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 690.711255] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777ab27d-64e2-49f3-a92e-d3c385eb506f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.854994] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794210, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062385} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.855320] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 690.856119] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa6ed6b-9621-4cbc-a282-ce031bbfde71 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.888465] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] d5a54c46-1a45-4bd5-8694-0defc757c455/d5a54c46-1a45-4bd5-8694-0defc757c455.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 690.888916] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a09ed9e5-a0ac-4e76-b766-fec55b400232 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.909197] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 690.909197] env[61995]: value = "task-794211" [ 690.909197] env[61995]: _type = "Task" [ 690.909197] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.918279] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794211, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.022851] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.084140] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.159402] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52962769-5ea3-688f-21e9-01e1d8305ac5, 'name': SearchDatastore_Task, 'duration_secs': 0.008115} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.160666] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2529b9dc-58d5-4e4d-917f-21f2d62b4809 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.166945] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 691.166945] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b7dbef-c040-e99e-0696-5ef154485c91" [ 691.166945] env[61995]: _type = "Task" [ 691.166945] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.176625] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b7dbef-c040-e99e-0696-5ef154485c91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.230349] env[61995]: INFO nova.compute.manager [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Took 29.33 seconds to build instance. [ 691.298596] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.299221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.421391] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794211, 'name': ReconfigVM_Task, 'duration_secs': 0.275808} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.421744] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Reconfigured VM instance instance-00000017 to attach disk [datastore1] d5a54c46-1a45-4bd5-8694-0defc757c455/d5a54c46-1a45-4bd5-8694-0defc757c455.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 691.422390] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46db3a8c-66d9-4ae8-b7d2-39ca2ea4c351 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.431959] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 691.431959] env[61995]: value = "task-794212" [ 691.431959] env[61995]: _type = "Task" [ 691.431959] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.445675] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794212, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.589957] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Releasing lock "refresh_cache-9f5024ce-5a48-4d27-a412-35149ee843f3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.590311] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61995) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.590433] env[61995]: DEBUG nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.590636] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.620460] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.683754] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b7dbef-c040-e99e-0696-5ef154485c91, 'name': SearchDatastore_Task, 'duration_secs': 0.009199} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.684584] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.684584] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] e78f8718-e0be-4b2a-ab70-9565ce01ab9d/e78f8718-e0be-4b2a-ab70-9565ce01ab9d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 691.684584] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ca1337f-134a-4b96-b724-be59719edb4f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.695938] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 691.695938] env[61995]: value = "task-794213" [ 691.695938] env[61995]: _type = "Task" [ 691.695938] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.707249] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794213, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.735763] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3294340d-70bc-4092-890d-e4c0f1c04e00 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.810s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.944292] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794212, 'name': Rename_Task, 'duration_secs': 0.163766} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.944717] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 691.944934] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d02c3316-4cca-4ac0-a5ff-2082cb329c83 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.952209] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 691.952209] env[61995]: value = "task-794214" [ 691.952209] env[61995]: _type = "Task" [ 691.952209] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.964193] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794214, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.025704] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30477022-789e-48d1-b768-16642df1199a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.034729] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4e588b-1d70-46f4-88f6-35846f884509 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.073799] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acde56c-c882-49c4-b4db-8534032e20b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.083016] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdee5144-35c4-4999-9762-173c7405936a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.099144] env[61995]: DEBUG nova.compute.provider_tree [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.128532] env[61995]: DEBUG nova.network.neutron [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.210139] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794213, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467844} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.210139] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] e78f8718-e0be-4b2a-ab70-9565ce01ab9d/e78f8718-e0be-4b2a-ab70-9565ce01ab9d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 692.210139] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 692.210139] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0aec045-8ff6-4d38-b093-0ee95d09773e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.216306] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 692.216306] env[61995]: value = "task-794215" [ 692.216306] env[61995]: _type = "Task" [ 692.216306] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.225744] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794215, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.244281] env[61995]: DEBUG nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.469694] env[61995]: DEBUG oslo_vmware.api [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794214, 'name': PowerOnVM_Task, 'duration_secs': 0.513993} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.470397] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 692.471574] env[61995]: INFO nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Took 7.48 seconds to spawn the instance on the hypervisor. [ 692.471954] env[61995]: DEBUG nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 692.472886] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cffbfb1-7fa3-4bcf-a6a7-2682718c5b33 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.602769] env[61995]: DEBUG nova.scheduler.client.report [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.630953] env[61995]: INFO nova.compute.manager [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: 9f5024ce-5a48-4d27-a412-35149ee843f3] Took 1.04 seconds to deallocate network for instance. [ 692.698286] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "06496d6a-324d-41d4-b3e5-52e3d93636bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.698602] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.698771] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "06496d6a-324d-41d4-b3e5-52e3d93636bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.698953] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.699143] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.701285] env[61995]: INFO nova.compute.manager [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Terminating instance [ 692.703138] env[61995]: DEBUG nova.compute.manager [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 692.703473] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 692.704400] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6834eed1-40d0-4b29-987a-a438dc737ab3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.712548] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 692.712802] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d92aeaa6-c801-470a-8721-b7edefc56dc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.721581] env[61995]: DEBUG oslo_vmware.api [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 692.721581] env[61995]: value = "task-794216" [ 692.721581] env[61995]: _type = "Task" [ 692.721581] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.727771] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794215, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070371} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.728393] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 692.729247] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305cfdd9-bf52-4f93-bc4b-a1df87e78451 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.734658] env[61995]: DEBUG oslo_vmware.api [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794216, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.754261] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] e78f8718-e0be-4b2a-ab70-9565ce01ab9d/e78f8718-e0be-4b2a-ab70-9565ce01ab9d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.756777] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db9a9649-1a0f-442c-82ad-a7d8ba57cff9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.779592] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 692.779592] env[61995]: value = "task-794217" [ 692.779592] env[61995]: _type = "Task" [ 692.779592] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.787527] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794217, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.790582] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.002606] env[61995]: INFO nova.compute.manager [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Took 29.15 seconds to build instance. [ 693.110440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.110440] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.111358] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.461s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.111683] env[61995]: DEBUG nova.objects.instance [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lazy-loading 'resources' on Instance uuid 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 693.236785] env[61995]: DEBUG oslo_vmware.api [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794216, 'name': PowerOffVM_Task, 'duration_secs': 0.276519} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.236861] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 693.237124] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 693.237458] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-970d6d75-0472-4e5e-8f3c-ea7d86917a3d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.292924] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794217, 'name': ReconfigVM_Task, 'duration_secs': 0.45342} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.292924] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Reconfigured VM instance instance-00000018 to attach disk [datastore1] e78f8718-e0be-4b2a-ab70-9565ce01ab9d/e78f8718-e0be-4b2a-ab70-9565ce01ab9d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 693.292924] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bdef0dd-348d-421c-8ef7-d14c14aa52b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.297953] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 693.297953] env[61995]: value = "task-794219" [ 693.297953] env[61995]: _type = "Task" [ 693.297953] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.299063] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 693.299257] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 693.299441] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Deleting the datastore file [datastore1] 06496d6a-324d-41d4-b3e5-52e3d93636bb {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 693.305876] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-545b6e5b-2d2a-4278-98d1-272745b92762 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.314235] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794219, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.315049] env[61995]: DEBUG oslo_vmware.api [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for the task: (returnval){ [ 693.315049] env[61995]: value = "task-794220" [ 693.315049] env[61995]: _type = "Task" [ 693.315049] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.322362] env[61995]: DEBUG oslo_vmware.api [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794220, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.503373] env[61995]: DEBUG oslo_concurrency.lockutils [None req-caee1207-3950-4cb6-88e8-0ed9dfd9ddef tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.045s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.614826] env[61995]: DEBUG nova.compute.utils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.619975] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.619975] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.667696] env[61995]: INFO nova.scheduler.client.report [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Deleted allocations for instance 9f5024ce-5a48-4d27-a412-35149ee843f3 [ 693.752131] env[61995]: DEBUG nova.policy [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3709c26007e499b8361e3c87075baa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '496222f4ed8a400b99044de6e97c439c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 693.811723] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794219, 'name': Rename_Task, 'duration_secs': 0.145526} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.814473] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 693.814816] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1adcf318-b5aa-4b9e-b9ee-811a8364662f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.824372] env[61995]: DEBUG oslo_vmware.api [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Task: {'id': task-794220, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190089} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.826084] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 693.826276] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 693.826452] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 693.826623] env[61995]: INFO nova.compute.manager [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 693.826857] env[61995]: DEBUG oslo.service.loopingcall [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.827111] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 693.827111] env[61995]: value = "task-794221" [ 693.827111] env[61995]: _type = "Task" [ 693.827111] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.827297] env[61995]: DEBUG nova.compute.manager [-] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.827393] env[61995]: DEBUG nova.network.neutron [-] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 693.839585] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.006291] env[61995]: DEBUG nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.073979] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880e85d6-aadf-4bed-9ec8-450f2f4d1d43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.081739] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61128af-3b38-429a-89a3-6c36cacb437d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.120756] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.126087] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd699b34-04ef-4f5c-a0f3-892d60f6cfda {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.133683] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14162ad5-d1c8-447a-b858-3da037874c61 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.149473] env[61995]: DEBUG nova.compute.provider_tree [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.189031] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0479cde3-d90f-4fce-8f1d-9fd04aa5566a tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "9f5024ce-5a48-4d27-a412-35149ee843f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.075s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.339292] env[61995]: DEBUG oslo_vmware.api [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794221, 'name': PowerOnVM_Task, 'duration_secs': 0.455373} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.339628] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 694.339940] env[61995]: INFO nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Took 4.87 seconds to spawn the instance on the hypervisor. [ 694.340063] env[61995]: DEBUG nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.341482] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e07f1d-dffa-410b-a4b6-eba751cbafe0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.446187] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Successfully created port: 7e64a241-d943-4260-9058-75a25d87a679 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.450512] env[61995]: DEBUG nova.compute.manager [req-7b3c53f9-4e6d-4672-ba0c-55853de1a913 req-f74a2ccf-9d99-4832-9e8f-22f0afea2829 service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Received event network-vif-deleted-c0c4eccc-870f-40e6-bbe0-f8c11172e6df {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 694.450512] env[61995]: INFO nova.compute.manager [req-7b3c53f9-4e6d-4672-ba0c-55853de1a913 req-f74a2ccf-9d99-4832-9e8f-22f0afea2829 service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Neutron deleted interface c0c4eccc-870f-40e6-bbe0-f8c11172e6df; detaching it from the instance and deleting it from the info cache [ 694.450629] env[61995]: DEBUG nova.network.neutron [req-7b3c53f9-4e6d-4672-ba0c-55853de1a913 req-f74a2ccf-9d99-4832-9e8f-22f0afea2829 service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.533357] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.651846] env[61995]: DEBUG nova.scheduler.client.report [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.691624] env[61995]: DEBUG nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.803279] env[61995]: DEBUG nova.compute.manager [None req-cab27d61-c548-4a9e-9097-513b6f1a3f27 tempest-ServerDiagnosticsV248Test-1604268663 tempest-ServerDiagnosticsV248Test-1604268663-project-admin] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.804416] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807b80e1-53dc-458f-95c3-8d8f1e37c563 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.812682] env[61995]: INFO nova.compute.manager [None req-cab27d61-c548-4a9e-9097-513b6f1a3f27 tempest-ServerDiagnosticsV248Test-1604268663 tempest-ServerDiagnosticsV248Test-1604268663-project-admin] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Retrieving diagnostics [ 694.813498] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d3da05-82fc-4413-b3ac-578ae7bb602c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.858432] env[61995]: INFO nova.compute.manager [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Took 26.34 seconds to build instance. [ 694.892959] env[61995]: DEBUG nova.network.neutron [-] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.953739] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f05895b-582c-44e9-8a39-814b3ee47619 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.963727] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6066530c-c93c-4682-828c-d021cb4bfe99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.986445] env[61995]: DEBUG nova.compute.manager [req-7b3c53f9-4e6d-4672-ba0c-55853de1a913 req-f74a2ccf-9d99-4832-9e8f-22f0afea2829 service nova] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Detach interface failed, port_id=c0c4eccc-870f-40e6-bbe0-f8c11172e6df, reason: Instance 06496d6a-324d-41d4-b3e5-52e3d93636bb could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 695.136585] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.156856] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.045s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.161199] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.207s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.162966] env[61995]: INFO nova.compute.claims [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.173827] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.174081] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.174236] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.174409] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.174548] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.174691] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.174893] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.175100] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.175355] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.175525] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.175688] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.176605] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81246313-88a2-47c7-b1b7-a139e5214739 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.185776] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef9d51d-a10a-4bd5-b2ef-1382fab61b9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.192892] env[61995]: INFO nova.scheduler.client.report [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Deleted allocations for instance 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc [ 695.232787] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.361836] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18de3737-3b8f-4dc0-8cfb-cf6093f4debd tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.391s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.396408] env[61995]: INFO nova.compute.manager [-] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Took 1.57 seconds to deallocate network for instance. [ 695.701419] env[61995]: DEBUG oslo_concurrency.lockutils [None req-df9f943a-2dc4-41aa-8ade-01f354ae5f34 tempest-ServersAdmin275Test-464003486 tempest-ServersAdmin275Test-464003486-project-member] Lock "2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.765s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.865684] env[61995]: DEBUG nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.915558] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "d5a54c46-1a45-4bd5-8694-0defc757c455" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.915558] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.915558] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "d5a54c46-1a45-4bd5-8694-0defc757c455-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.915558] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.915763] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.919195] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.919811] env[61995]: INFO nova.compute.manager [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Terminating instance [ 695.921877] env[61995]: DEBUG nova.compute.manager [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.922509] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 695.923610] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3929f2-1cdb-41b7-82f4-a7ca51d0e3f3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.934037] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 695.934037] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f668cd9-c7ea-4748-971f-0b5324043b56 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.945173] env[61995]: DEBUG oslo_vmware.api [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 695.945173] env[61995]: value = "task-794222" [ 695.945173] env[61995]: _type = "Task" [ 695.945173] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.954060] env[61995]: DEBUG oslo_vmware.api [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794222, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.393652] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.445301] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Successfully updated port: 7e64a241-d943-4260-9058-75a25d87a679 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 696.457706] env[61995]: DEBUG oslo_vmware.api [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794222, 'name': PowerOffVM_Task, 'duration_secs': 0.204599} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.462137] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 696.462137] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 696.463403] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2fa2d1c-8a9c-4913-a1be-9666ae083686 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.482952] env[61995]: DEBUG nova.compute.manager [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Received event network-vif-plugged-7e64a241-d943-4260-9058-75a25d87a679 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 696.484663] env[61995]: DEBUG oslo_concurrency.lockutils [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] Acquiring lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.484663] env[61995]: DEBUG oslo_concurrency.lockutils [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.484663] env[61995]: DEBUG oslo_concurrency.lockutils [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.484663] env[61995]: DEBUG nova.compute.manager [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] No waiting events found dispatching network-vif-plugged-7e64a241-d943-4260-9058-75a25d87a679 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 696.484663] env[61995]: WARNING nova.compute.manager [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Received unexpected event network-vif-plugged-7e64a241-d943-4260-9058-75a25d87a679 for instance with vm_state building and task_state spawning. [ 696.484819] env[61995]: DEBUG nova.compute.manager [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Received event network-changed-7e64a241-d943-4260-9058-75a25d87a679 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 696.484819] env[61995]: DEBUG nova.compute.manager [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Refreshing instance network info cache due to event network-changed-7e64a241-d943-4260-9058-75a25d87a679. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 696.484819] env[61995]: DEBUG oslo_concurrency.lockutils [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] Acquiring lock "refresh_cache-f47e5bbb-f4a8-4686-8195-8348d2ffb08d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.484907] env[61995]: DEBUG oslo_concurrency.lockutils [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] Acquired lock "refresh_cache-f47e5bbb-f4a8-4686-8195-8348d2ffb08d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.485819] env[61995]: DEBUG nova.network.neutron [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Refreshing network info cache for port 7e64a241-d943-4260-9058-75a25d87a679 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 696.531512] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 696.531734] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 696.531935] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Deleting the datastore file [datastore1] d5a54c46-1a45-4bd5-8694-0defc757c455 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 696.532189] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e67562d-e961-4d60-8b2b-83a27d89499b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.540654] env[61995]: DEBUG oslo_vmware.api [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for the task: (returnval){ [ 696.540654] env[61995]: value = "task-794224" [ 696.540654] env[61995]: _type = "Task" [ 696.540654] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.553988] env[61995]: DEBUG oslo_vmware.api [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794224, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.564421] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b688ca90-bd7b-4dcd-a95b-ce43acb38334 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.573336] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487c3594-c802-4f8f-a9c6-28193ffcf2c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.601584] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d72c7cc-ae3e-456f-865e-340eac4c4ab1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.609504] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0180e7e9-0f26-47d7-bb3e-af558b081a57 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.623713] env[61995]: DEBUG nova.compute.provider_tree [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.949020] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "refresh_cache-f47e5bbb-f4a8-4686-8195-8348d2ffb08d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.053433] env[61995]: DEBUG oslo_vmware.api [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Task: {'id': task-794224, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265152} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.054305] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 697.054587] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 697.054847] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 697.055121] env[61995]: INFO nova.compute.manager [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Took 1.13 seconds to destroy the instance on the hypervisor. [ 697.055412] env[61995]: DEBUG oslo.service.loopingcall [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.055656] env[61995]: DEBUG nova.compute.manager [-] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.055788] env[61995]: DEBUG nova.network.neutron [-] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.109554] env[61995]: DEBUG nova.network.neutron [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.126920] env[61995]: DEBUG nova.scheduler.client.report [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.261536] env[61995]: DEBUG nova.network.neutron [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.634939] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.635484] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.638889] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.205s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.640276] env[61995]: INFO nova.compute.claims [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.763779] env[61995]: DEBUG oslo_concurrency.lockutils [req-26c6a882-cceb-4ca5-88fb-405b31800488 req-3b3f68f6-325d-4208-9806-27b3540854f6 service nova] Releasing lock "refresh_cache-f47e5bbb-f4a8-4686-8195-8348d2ffb08d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.764220] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "refresh_cache-f47e5bbb-f4a8-4686-8195-8348d2ffb08d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.764385] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.932251] env[61995]: DEBUG nova.network.neutron [-] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.145472] env[61995]: DEBUG nova.compute.utils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.149536] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 698.149818] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 698.270511] env[61995]: DEBUG nova.policy [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3709c26007e499b8361e3c87075baa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '496222f4ed8a400b99044de6e97c439c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 698.345579] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.437499] env[61995]: INFO nova.compute.manager [-] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Took 1.38 seconds to deallocate network for instance. [ 698.645955] env[61995]: DEBUG nova.compute.manager [req-8af06d39-85d4-4713-b13b-bd91702fb3e4 req-8014a709-d05b-4757-a2c4-b7fd169b5cca service nova] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Received event network-vif-deleted-b1c0d4ae-5394-41b6-a0fb-1e990a1ed6c0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 698.650016] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.739356] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Updating instance_info_cache with network_info: [{"id": "7e64a241-d943-4260-9058-75a25d87a679", "address": "fa:16:3e:ae:47:9d", "network": {"id": "a40cc2c5-85a3-4ff0-a6c0-36f98fc96249", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-783559288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "496222f4ed8a400b99044de6e97c439c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e64a241-d9", "ovs_interfaceid": "7e64a241-d943-4260-9058-75a25d87a679", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.946392] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.969530] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Successfully created port: 71f99164-4e02-4b29-bb4c-7a576848a8ee {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.060706] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4927901-4b51-462b-bffe-c8c88f0fe4f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.069506] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d4db32-cfd0-4ea7-9ba3-55feb2a04d13 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.109871] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a27c87-202c-454f-a1a3-da79b2778eb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.115688] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.115688] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.121181] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4145d6-bec3-4ed8-a3f0-6817928010ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.134937] env[61995]: DEBUG nova.compute.provider_tree [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.241788] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "refresh_cache-f47e5bbb-f4a8-4686-8195-8348d2ffb08d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.242137] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Instance network_info: |[{"id": "7e64a241-d943-4260-9058-75a25d87a679", "address": "fa:16:3e:ae:47:9d", "network": {"id": "a40cc2c5-85a3-4ff0-a6c0-36f98fc96249", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-783559288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "496222f4ed8a400b99044de6e97c439c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e64a241-d9", "ovs_interfaceid": "7e64a241-d943-4260-9058-75a25d87a679", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 699.242561] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:47:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '877ed63d-906e-4bd5-a1fc-7e82d172d41e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e64a241-d943-4260-9058-75a25d87a679', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.251924] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Creating folder: Project (496222f4ed8a400b99044de6e97c439c). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 699.252779] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aee9881c-8e7d-4460-823a-bb0929be9767 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.264729] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Created folder: Project (496222f4ed8a400b99044de6e97c439c) in parent group-v185203. [ 699.264915] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Creating folder: Instances. Parent ref: group-v185218. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 699.265164] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96864bc1-8d54-4464-aa72-b45f3edbc486 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.275190] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Created folder: Instances in parent group-v185218. [ 699.275439] env[61995]: DEBUG oslo.service.loopingcall [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.276402] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 699.276402] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53ed159b-a2e5-470a-ac9e-5e76cac45401 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.294679] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.294679] env[61995]: value = "task-794227" [ 699.294679] env[61995]: _type = "Task" [ 699.294679] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.302655] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794227, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.640341] env[61995]: DEBUG nova.scheduler.client.report [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.668264] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.702891] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.702891] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.702891] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.703101] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.703101] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.703101] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.703101] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.703311] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.703489] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.703880] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.704125] env[61995]: DEBUG nova.virt.hardware [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.705572] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a45300b-9f0b-4f27-bbbc-64870b42b262 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.721091] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa5c74a-20ef-4b8d-b920-a40641182b45 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.807638] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794227, 'name': CreateVM_Task, 'duration_secs': 0.334228} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.809694] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 699.809694] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.809694] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.809694] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 699.809694] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd854710-9b88-475d-b830-ec44544e467e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.815544] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 699.815544] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52af6619-613b-760d-ae24-aa5346e4be40" [ 699.815544] env[61995]: _type = "Task" [ 699.815544] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.827723] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52af6619-613b-760d-ae24-aa5346e4be40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.150875] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.150875] env[61995]: DEBUG nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.153139] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.660s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.154890] env[61995]: INFO nova.compute.claims [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.327996] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52af6619-613b-760d-ae24-aa5346e4be40, 'name': SearchDatastore_Task, 'duration_secs': 0.047683} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.331996] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.332938] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.333271] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.333428] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.333608] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.337223] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7621dc3-46c0-4aa6-bcb0-edf9859be8d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.346730] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.346915] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 700.347621] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e046915-e5c9-4260-8b0a-7378a336604d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.353293] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 700.353293] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5294f237-07a4-b032-8b22-c1d543e1cb34" [ 700.353293] env[61995]: _type = "Task" [ 700.353293] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.361132] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5294f237-07a4-b032-8b22-c1d543e1cb34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.659841] env[61995]: DEBUG nova.compute.utils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.663418] env[61995]: DEBUG nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.663502] env[61995]: DEBUG nova.network.neutron [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 700.760330] env[61995]: DEBUG nova.policy [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fda120058c42481f9b06de4b03539f44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df7381568d0e44fbb8bb161e9d5e8bb7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 700.864165] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5294f237-07a4-b032-8b22-c1d543e1cb34, 'name': SearchDatastore_Task, 'duration_secs': 0.014865} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.865640] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8d45420-a441-4257-9367-b22aa5dbfba4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.871597] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 700.871597] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52068d86-c8b3-51b5-bea1-c9e88334bc51" [ 700.871597] env[61995]: _type = "Task" [ 700.871597] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.880885] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52068d86-c8b3-51b5-bea1-c9e88334bc51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.052782] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "eef1cf32-e71e-4696-8170-1023e3a5518d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.053020] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.074475] env[61995]: DEBUG nova.compute.manager [req-12a76c2b-9715-4dbf-84fc-793e83e47714 req-fd2685fe-b797-4982-9276-77577e6d8815 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Received event network-vif-plugged-71f99164-4e02-4b29-bb4c-7a576848a8ee {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 701.074688] env[61995]: DEBUG oslo_concurrency.lockutils [req-12a76c2b-9715-4dbf-84fc-793e83e47714 req-fd2685fe-b797-4982-9276-77577e6d8815 service nova] Acquiring lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.074889] env[61995]: DEBUG oslo_concurrency.lockutils [req-12a76c2b-9715-4dbf-84fc-793e83e47714 req-fd2685fe-b797-4982-9276-77577e6d8815 service nova] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.075062] env[61995]: DEBUG oslo_concurrency.lockutils [req-12a76c2b-9715-4dbf-84fc-793e83e47714 req-fd2685fe-b797-4982-9276-77577e6d8815 service nova] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.075232] env[61995]: DEBUG nova.compute.manager [req-12a76c2b-9715-4dbf-84fc-793e83e47714 req-fd2685fe-b797-4982-9276-77577e6d8815 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] No waiting events found dispatching network-vif-plugged-71f99164-4e02-4b29-bb4c-7a576848a8ee {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 701.075458] env[61995]: WARNING nova.compute.manager [req-12a76c2b-9715-4dbf-84fc-793e83e47714 req-fd2685fe-b797-4982-9276-77577e6d8815 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Received unexpected event network-vif-plugged-71f99164-4e02-4b29-bb4c-7a576848a8ee for instance with vm_state building and task_state spawning. [ 701.164921] env[61995]: DEBUG nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 701.337745] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Successfully updated port: 71f99164-4e02-4b29-bb4c-7a576848a8ee {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 701.371011] env[61995]: DEBUG nova.network.neutron [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Successfully created port: 52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.388801] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52068d86-c8b3-51b5-bea1-c9e88334bc51, 'name': SearchDatastore_Task, 'duration_secs': 0.019021} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.389556] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.389556] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] f47e5bbb-f4a8-4686-8195-8348d2ffb08d/f47e5bbb-f4a8-4686-8195-8348d2ffb08d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 701.396142] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-658a44b6-41d3-450f-9635-21be88a5c2b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.405156] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 701.405156] env[61995]: value = "task-794228" [ 701.405156] env[61995]: _type = "Task" [ 701.405156] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.411414] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.614923] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b52e138-ecc9-4847-bc9b-9a43c2c5c616 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.624424] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3cb8e5-6328-4c20-b432-d39030c09d8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.666560] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d6f885-b401-49ec-8213-a6faa6833cb1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.682588] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b1401e-b9ea-4e57-9d40-71c795422eb4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.701813] env[61995]: DEBUG nova.compute.provider_tree [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.840472] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "refresh_cache-08489e01-2ce3-45a1-8f87-2a4c89230aa2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.841529] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "refresh_cache-08489e01-2ce3-45a1-8f87-2a4c89230aa2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.841529] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 701.919025] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794228, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43779} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.919025] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] f47e5bbb-f4a8-4686-8195-8348d2ffb08d/f47e5bbb-f4a8-4686-8195-8348d2ffb08d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 701.919025] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 701.919025] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0583f53f-7871-4a42-8ee1-6aa4b391cb20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.926024] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 701.926024] env[61995]: value = "task-794229" [ 701.926024] env[61995]: _type = "Task" [ 701.926024] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.935124] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794229, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.182213] env[61995]: DEBUG nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.206356] env[61995]: DEBUG nova.scheduler.client.report [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.218336] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.218336] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.218511] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.218682] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.218814] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.218992] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.220086] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.221226] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.221226] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.221226] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.221226] env[61995]: DEBUG nova.virt.hardware [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.224245] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d70c7e-86a6-45e6-828d-b356398a140f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.235599] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c82692-6db5-4478-be81-22973a92750a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.402615] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.436625] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794229, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127739} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.436890] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 702.437668] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabe0569-3814-4cac-a720-746f6b44ac56 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.462899] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] f47e5bbb-f4a8-4686-8195-8348d2ffb08d/f47e5bbb-f4a8-4686-8195-8348d2ffb08d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 702.463420] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdd97958-9200-43b1-9703-d8b65e01e912 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.490864] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 702.490864] env[61995]: value = "task-794230" [ 702.490864] env[61995]: _type = "Task" [ 702.490864] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.498648] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794230, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.618139] env[61995]: DEBUG nova.network.neutron [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Updating instance_info_cache with network_info: [{"id": "71f99164-4e02-4b29-bb4c-7a576848a8ee", "address": "fa:16:3e:1b:8d:5d", "network": {"id": "a40cc2c5-85a3-4ff0-a6c0-36f98fc96249", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-783559288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "496222f4ed8a400b99044de6e97c439c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71f99164-4e", "ovs_interfaceid": "71f99164-4e02-4b29-bb4c-7a576848a8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.712696] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.713666] env[61995]: DEBUG nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 702.716349] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.024s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.717620] env[61995]: INFO nova.compute.claims [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.000571] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794230, 'name': ReconfigVM_Task, 'duration_secs': 0.264522} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.003319] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Reconfigured VM instance instance-00000019 to attach disk [datastore2] f47e5bbb-f4a8-4686-8195-8348d2ffb08d/f47e5bbb-f4a8-4686-8195-8348d2ffb08d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 703.003319] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8840d1a5-3c25-411c-9ce5-6bb23b9a2847 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.012018] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 703.012018] env[61995]: value = "task-794231" [ 703.012018] env[61995]: _type = "Task" [ 703.012018] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.022366] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794231, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.051138] env[61995]: DEBUG nova.network.neutron [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Successfully updated port: 52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 703.121010] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "refresh_cache-08489e01-2ce3-45a1-8f87-2a4c89230aa2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.121362] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Instance network_info: |[{"id": "71f99164-4e02-4b29-bb4c-7a576848a8ee", "address": "fa:16:3e:1b:8d:5d", "network": {"id": "a40cc2c5-85a3-4ff0-a6c0-36f98fc96249", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-783559288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "496222f4ed8a400b99044de6e97c439c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71f99164-4e", "ovs_interfaceid": "71f99164-4e02-4b29-bb4c-7a576848a8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 703.121774] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:8d:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '877ed63d-906e-4bd5-a1fc-7e82d172d41e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71f99164-4e02-4b29-bb4c-7a576848a8ee', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 703.130320] env[61995]: DEBUG oslo.service.loopingcall [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.130945] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 703.131206] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa08f633-24aa-423d-af60-77609995c697 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.150178] env[61995]: DEBUG nova.compute.manager [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Received event network-changed-71f99164-4e02-4b29-bb4c-7a576848a8ee {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 703.150419] env[61995]: DEBUG nova.compute.manager [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Refreshing instance network info cache due to event network-changed-71f99164-4e02-4b29-bb4c-7a576848a8ee. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 703.150596] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Acquiring lock "refresh_cache-08489e01-2ce3-45a1-8f87-2a4c89230aa2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.150990] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Acquired lock "refresh_cache-08489e01-2ce3-45a1-8f87-2a4c89230aa2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.150990] env[61995]: DEBUG nova.network.neutron [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Refreshing network info cache for port 71f99164-4e02-4b29-bb4c-7a576848a8ee {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 703.155201] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 703.155201] env[61995]: value = "task-794232" [ 703.155201] env[61995]: _type = "Task" [ 703.155201] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.164427] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794232, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.222146] env[61995]: DEBUG nova.compute.utils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.223860] env[61995]: DEBUG nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.224458] env[61995]: DEBUG nova.network.neutron [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 703.270226] env[61995]: DEBUG nova.policy [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e08719d59b456786fce984cc1508e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c3e03b3825c43ca987d60b339fc0a52', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 703.520597] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794231, 'name': Rename_Task, 'duration_secs': 0.130688} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.520890] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 703.521160] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9c68bce-1362-42cd-89e2-49057ffb9eb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.527751] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 703.527751] env[61995]: value = "task-794233" [ 703.527751] env[61995]: _type = "Task" [ 703.527751] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.536438] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.557146] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.557294] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.557446] env[61995]: DEBUG nova.network.neutron [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.664990] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794232, 'name': CreateVM_Task, 'duration_secs': 0.332358} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.665182] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 703.665878] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.666061] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.666331] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 703.666576] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b572f17-b7bb-47af-9db3-a030ad7feff3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.671378] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 703.671378] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f33a1b-df2a-7ed6-1e9e-6b56f940ba15" [ 703.671378] env[61995]: _type = "Task" [ 703.671378] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.679343] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f33a1b-df2a-7ed6-1e9e-6b56f940ba15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.727271] env[61995]: DEBUG nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.837514] env[61995]: DEBUG nova.network.neutron [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Successfully created port: 802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.050058] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794233, 'name': PowerOnVM_Task, 'duration_secs': 0.454943} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.052558] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 704.052770] env[61995]: INFO nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Took 8.92 seconds to spawn the instance on the hypervisor. [ 704.052946] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 704.054118] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb52cb27-4553-44e9-96f3-34e5fc582c89 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.115963] env[61995]: DEBUG nova.network.neutron [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Updated VIF entry in instance network info cache for port 71f99164-4e02-4b29-bb4c-7a576848a8ee. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 704.115963] env[61995]: DEBUG nova.network.neutron [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Updating instance_info_cache with network_info: [{"id": "71f99164-4e02-4b29-bb4c-7a576848a8ee", "address": "fa:16:3e:1b:8d:5d", "network": {"id": "a40cc2c5-85a3-4ff0-a6c0-36f98fc96249", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-783559288-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "496222f4ed8a400b99044de6e97c439c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71f99164-4e", "ovs_interfaceid": "71f99164-4e02-4b29-bb4c-7a576848a8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.128772] env[61995]: DEBUG nova.network.neutron [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.157277] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5d1467-1c57-4645-bd3e-0d6cf1602612 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.167790] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf2d733-59df-4b4a-ba6d-ad6f48d6292b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.208586] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e24416-f019-4d81-83f9-b89bd7a5f80a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.218413] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f33a1b-df2a-7ed6-1e9e-6b56f940ba15, 'name': SearchDatastore_Task, 'duration_secs': 0.008917} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.220035] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.220035] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.220035] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.220035] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.220210] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.220210] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b47b3466-c8b1-432e-8b62-add200639002 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.224693] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e5b532-feed-479b-bfa6-6af58faf3024 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.229472] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.229650] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 704.230677] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ba7e228-b3ae-4eb3-a22c-f48450512220 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.245388] env[61995]: DEBUG nova.compute.provider_tree [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.249485] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 704.249485] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529f26c9-c06e-5661-3191-4ac1d2cd1ece" [ 704.249485] env[61995]: _type = "Task" [ 704.249485] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.256944] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529f26c9-c06e-5661-3191-4ac1d2cd1ece, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.392836] env[61995]: DEBUG nova.network.neutron [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.577517] env[61995]: INFO nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Took 31.51 seconds to build instance. [ 704.618575] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Releasing lock "refresh_cache-08489e01-2ce3-45a1-8f87-2a4c89230aa2" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.618846] env[61995]: DEBUG nova.compute.manager [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received event network-vif-plugged-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 704.619169] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Acquiring lock "e53d10a3-1d16-498d-b4d8-abe7205410ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.619382] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.619689] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.619763] env[61995]: DEBUG nova.compute.manager [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] No waiting events found dispatching network-vif-plugged-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 704.619865] env[61995]: WARNING nova.compute.manager [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received unexpected event network-vif-plugged-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 for instance with vm_state building and task_state spawning. [ 704.620034] env[61995]: DEBUG nova.compute.manager [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 704.620376] env[61995]: DEBUG nova.compute.manager [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing instance network info cache due to event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 704.620611] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Acquiring lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.748020] env[61995]: DEBUG nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.750035] env[61995]: DEBUG nova.scheduler.client.report [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.775649] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529f26c9-c06e-5661-3191-4ac1d2cd1ece, 'name': SearchDatastore_Task, 'duration_secs': 0.008575} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.779754] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-305e5cb4-36d6-4eb1-936d-015c93da873b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.789020] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 704.789020] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f707df-c0d0-e899-5dcd-5c549b8d6e95" [ 704.789020] env[61995]: _type = "Task" [ 704.789020] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.789600] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.789969] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.790304] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.790606] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.790874] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.791192] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.791513] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.791775] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.792060] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.792753] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.793256] env[61995]: DEBUG nova.virt.hardware [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.795254] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f6cdc7-d701-4a08-ab1d-e3703f13ff65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.809987] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd218b3-e164-4ccd-bb67-3b302f1c2347 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.813696] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f707df-c0d0-e899-5dcd-5c549b8d6e95, 'name': SearchDatastore_Task, 'duration_secs': 0.00856} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.814081] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.814440] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 08489e01-2ce3-45a1-8f87-2a4c89230aa2/08489e01-2ce3-45a1-8f87-2a4c89230aa2.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 704.815125] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0a72c1f-4d99-4d38-b8d9-755e74ccf763 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.832020] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 704.832020] env[61995]: value = "task-794234" [ 704.832020] env[61995]: _type = "Task" [ 704.832020] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.838929] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.895907] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Releasing lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.896318] env[61995]: DEBUG nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Instance network_info: |[{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 704.896728] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Acquired lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.897017] env[61995]: DEBUG nova.network.neutron [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 704.898318] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:bb:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '352165bb-004f-4180-9627-3a275dbe18af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52c15b2e-a283-4a39-9a38-d2ccd7da93a2', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 704.908350] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Creating folder: Project (df7381568d0e44fbb8bb161e9d5e8bb7). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 704.911895] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a02a6e67-88f8-4791-a34c-4704e3e9e568 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.924109] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Created folder: Project (df7381568d0e44fbb8bb161e9d5e8bb7) in parent group-v185203. [ 704.924109] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Creating folder: Instances. Parent ref: group-v185222. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 704.924109] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c595e9df-198c-4590-9471-de7fcb3fa170 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.933735] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Created folder: Instances in parent group-v185222. [ 704.934063] env[61995]: DEBUG oslo.service.loopingcall [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.934235] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 704.934405] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef043919-d729-4b6f-886e-43cc9c78154d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.956942] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.956942] env[61995]: value = "task-794237" [ 704.956942] env[61995]: _type = "Task" [ 704.956942] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.965071] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794237, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.080483] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.952s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.267031] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.267031] env[61995]: DEBUG nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.271177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.647s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.271692] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.272047] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 705.272452] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.409s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.274680] env[61995]: INFO nova.compute.claims [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.279213] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f366d75-6526-424d-a9e7-0ec19c2ac53f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.294592] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ff8c40-f904-45fa-84aa-f34890b94bae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.318186] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c7e177-73f2-4d58-92a7-61bd3e323457 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.328158] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32347434-9f18-4e9b-ad46-26374e36e322 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.342397] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794234, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477032} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.366024] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 08489e01-2ce3-45a1-8f87-2a4c89230aa2/08489e01-2ce3-45a1-8f87-2a4c89230aa2.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 705.366342] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 705.367082] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181659MB free_disk=67GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 705.367266] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.367635] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9acd6d86-f675-45eb-8223-e499c5d2857c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.376103] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 705.376103] env[61995]: value = "task-794238" [ 705.376103] env[61995]: _type = "Task" [ 705.376103] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.378904] env[61995]: DEBUG nova.compute.manager [None req-f4ea2ddd-c469-46d5-8d1d-febe8af1ff7e tempest-ServerDiagnosticsV248Test-1604268663 tempest-ServerDiagnosticsV248Test-1604268663-project-admin] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 705.386483] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c344d15-b85e-4cb3-b70c-92b080ebff38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.394486] env[61995]: INFO nova.compute.manager [None req-f4ea2ddd-c469-46d5-8d1d-febe8af1ff7e tempest-ServerDiagnosticsV248Test-1604268663 tempest-ServerDiagnosticsV248Test-1604268663-project-admin] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Retrieving diagnostics [ 705.399144] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c409434-fc61-439d-ae4f-5ac4363e2c36 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.401968] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794238, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.465587] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794237, 'name': CreateVM_Task, 'duration_secs': 0.464966} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.465978] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 705.466725] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.466838] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.467262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 705.467848] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36189e7c-79c7-489e-ba42-9c3c28b7fbb7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.472453] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 705.472453] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527ab708-87bf-f4be-a282-01525103bf69" [ 705.472453] env[61995]: _type = "Task" [ 705.472453] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.480166] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527ab708-87bf-f4be-a282-01525103bf69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.582349] env[61995]: DEBUG nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 705.610225] env[61995]: DEBUG nova.network.neutron [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updated VIF entry in instance network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 705.610641] env[61995]: DEBUG nova.network.neutron [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.772809] env[61995]: DEBUG nova.compute.utils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.774209] env[61995]: DEBUG nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 705.774511] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 705.875270] env[61995]: DEBUG nova.policy [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9456ac4a7aa64ef09b256834206d2608', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f22192b1bd12430e9f2de0288d1e7e2a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 705.886543] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794238, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098111} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.886797] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 705.887671] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e971dea4-88a3-4886-97e5-419400ac8ae3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.909216] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] 08489e01-2ce3-45a1-8f87-2a4c89230aa2/08489e01-2ce3-45a1-8f87-2a4c89230aa2.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 705.909479] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1073d85-9100-4a3f-87df-48a366d385e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.928442] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 705.928442] env[61995]: value = "task-794239" [ 705.928442] env[61995]: _type = "Task" [ 705.928442] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.941158] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794239, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.985129] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527ab708-87bf-f4be-a282-01525103bf69, 'name': SearchDatastore_Task, 'duration_secs': 0.008143} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.985460] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.985685] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 705.985930] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.986100] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.986283] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 705.986542] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3888a0f-1e62-4ef2-adfb-b57689c581d1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.994230] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 705.994964] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 705.995111] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d83e6b69-87b7-4e60-a9d8-3dba8018315f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.000600] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 706.000600] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f0bccc-2370-66b6-d277-7b8e0b727e44" [ 706.000600] env[61995]: _type = "Task" [ 706.000600] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.008360] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f0bccc-2370-66b6-d277-7b8e0b727e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.026076] env[61995]: DEBUG nova.compute.manager [req-632943df-6381-42fd-a91d-aa69df805df5 req-f1ac671f-ace0-42e8-8d46-fca192378074 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Received event network-vif-plugged-802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 706.026319] env[61995]: DEBUG oslo_concurrency.lockutils [req-632943df-6381-42fd-a91d-aa69df805df5 req-f1ac671f-ace0-42e8-8d46-fca192378074 service nova] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.026521] env[61995]: DEBUG oslo_concurrency.lockutils [req-632943df-6381-42fd-a91d-aa69df805df5 req-f1ac671f-ace0-42e8-8d46-fca192378074 service nova] Lock "d9c53806-68c0-4872-a262-36bc05573674-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.026688] env[61995]: DEBUG oslo_concurrency.lockutils [req-632943df-6381-42fd-a91d-aa69df805df5 req-f1ac671f-ace0-42e8-8d46-fca192378074 service nova] Lock "d9c53806-68c0-4872-a262-36bc05573674-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.026850] env[61995]: DEBUG nova.compute.manager [req-632943df-6381-42fd-a91d-aa69df805df5 req-f1ac671f-ace0-42e8-8d46-fca192378074 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] No waiting events found dispatching network-vif-plugged-802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 706.027014] env[61995]: WARNING nova.compute.manager [req-632943df-6381-42fd-a91d-aa69df805df5 req-f1ac671f-ace0-42e8-8d46-fca192378074 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Received unexpected event network-vif-plugged-802b752a-238e-4b55-8bd3-f73eddc4b0f4 for instance with vm_state building and task_state spawning. [ 706.112761] env[61995]: DEBUG nova.network.neutron [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Successfully updated port: 802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 706.114960] env[61995]: DEBUG oslo_concurrency.lockutils [req-ff14eeec-f9b1-4e36-81b1-99a7fb61237f req-d6dcf1cb-e8f3-4169-b238-9df418a57b19 service nova] Releasing lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.116308] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.278419] env[61995]: DEBUG nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 706.371596] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Successfully created port: ebbd71fc-00dd-4b27-95ed-6670ac9b552e {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.443403] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794239, 'name': ReconfigVM_Task, 'duration_secs': 0.27657} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.443723] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Reconfigured VM instance instance-0000001a to attach disk [datastore2] 08489e01-2ce3-45a1-8f87-2a4c89230aa2/08489e01-2ce3-45a1-8f87-2a4c89230aa2.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 706.444829] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39dd6683-4ee8-4a42-ac2b-8c6cd7137585 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.452497] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 706.452497] env[61995]: value = "task-794240" [ 706.452497] env[61995]: _type = "Task" [ 706.452497] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.461061] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794240, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.511182] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f0bccc-2370-66b6-d277-7b8e0b727e44, 'name': SearchDatastore_Task, 'duration_secs': 0.008474} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.511792] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4a918e8-2ae8-4759-b3a3-6e6d7812d6e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.520103] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 706.520103] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5229cd87-c9aa-b001-4aba-4fe163a485c8" [ 706.520103] env[61995]: _type = "Task" [ 706.520103] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.530068] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5229cd87-c9aa-b001-4aba-4fe163a485c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.618253] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.618253] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.618253] env[61995]: DEBUG nova.network.neutron [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.720563] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.720820] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.721035] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.721217] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.721382] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.723675] env[61995]: INFO nova.compute.manager [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Terminating instance [ 706.726423] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "refresh_cache-e78f8718-e0be-4b2a-ab70-9565ce01ab9d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.726579] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquired lock "refresh_cache-e78f8718-e0be-4b2a-ab70-9565ce01ab9d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.726741] env[61995]: DEBUG nova.network.neutron [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.732055] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd01c4b-8a3d-4edc-a5c2-873a2f6e094e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.735282] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Successfully created port: da35a09f-ea6a-43b8-a817-8a2efbe499bb {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.742676] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39af31b9-f906-4bcc-b128-6fe359425700 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.774443] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d8002d-e2ab-41ef-ace7-42e03ade03f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.789021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32aa085-4b82-466c-88eb-69e86a7b21a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.801254] env[61995]: DEBUG nova.compute.provider_tree [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.963171] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794240, 'name': Rename_Task, 'duration_secs': 0.137389} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.963465] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 706.963741] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-626df52e-0540-499c-849d-1d6f0d388808 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.970991] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 706.970991] env[61995]: value = "task-794241" [ 706.970991] env[61995]: _type = "Task" [ 706.970991] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.978733] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794241, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.030706] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5229cd87-c9aa-b001-4aba-4fe163a485c8, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.033369] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.033369] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] e53d10a3-1d16-498d-b4d8-abe7205410ec/e53d10a3-1d16-498d-b4d8-abe7205410ec.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 707.033369] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da8c89c5-a275-443e-b7f0-a964bcdd3f3a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.034734] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Successfully created port: b569b424-48e7-45f4-8f0d-53192e53abf0 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.040857] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 707.040857] env[61995]: value = "task-794242" [ 707.040857] env[61995]: _type = "Task" [ 707.040857] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.049542] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794242, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.179788] env[61995]: DEBUG nova.network.neutron [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.258170] env[61995]: DEBUG nova.network.neutron [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.292190] env[61995]: DEBUG nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.305536] env[61995]: DEBUG nova.scheduler.client.report [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.331420] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.332122] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.332378] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.332847] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.333064] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.333275] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.333595] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.333787] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.334041] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.334258] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.334506] env[61995]: DEBUG nova.virt.hardware [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.336274] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b171f925-13eb-4571-8ba5-c0efd51411af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.351696] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2260b3d-9f9d-4fff-a90e-2c7847881b45 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.429394] env[61995]: DEBUG nova.network.neutron [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.482967] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794241, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.551028] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794242, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473116} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.551028] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] e53d10a3-1d16-498d-b4d8-abe7205410ec/e53d10a3-1d16-498d-b4d8-abe7205410ec.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 707.551028] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 707.551280] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55e7d593-2923-4e40-ab02-064693e2e036 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.558050] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 707.558050] env[61995]: value = "task-794243" [ 707.558050] env[61995]: _type = "Task" [ 707.558050] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.566426] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794243, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.654013] env[61995]: DEBUG nova.network.neutron [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.812112] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.812848] env[61995]: DEBUG nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 707.815548] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.025s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.817530] env[61995]: INFO nova.compute.claims [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.932439] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Releasing lock "refresh_cache-e78f8718-e0be-4b2a-ab70-9565ce01ab9d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.932886] env[61995]: DEBUG nova.compute.manager [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.933132] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 707.934708] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670a61bb-6034-4b9b-8422-04cdd715d1ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.942366] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 707.942931] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-444a2269-3149-4e01-b12d-df08aa676293 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.949103] env[61995]: DEBUG oslo_vmware.api [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 707.949103] env[61995]: value = "task-794244" [ 707.949103] env[61995]: _type = "Task" [ 707.949103] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.957043] env[61995]: DEBUG oslo_vmware.api [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.981500] env[61995]: DEBUG oslo_vmware.api [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794241, 'name': PowerOnVM_Task, 'duration_secs': 0.518002} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.981797] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 707.982029] env[61995]: INFO nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Took 8.32 seconds to spawn the instance on the hypervisor. [ 707.982239] env[61995]: DEBUG nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 707.983079] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3317a652-8ddb-4bf0-913f-9429b66e8c94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.068324] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794243, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182749} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.068681] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 708.069516] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123b7cb1-8571-4585-9a8d-fcb1b96b2b73 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.095236] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] e53d10a3-1d16-498d-b4d8-abe7205410ec/e53d10a3-1d16-498d-b4d8-abe7205410ec.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 708.095569] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6911f85c-41ee-4bf7-87cb-1c8f2236a0da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.115789] env[61995]: DEBUG nova.compute.manager [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Received event network-changed-802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 708.115991] env[61995]: DEBUG nova.compute.manager [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Refreshing instance network info cache due to event network-changed-802b752a-238e-4b55-8bd3-f73eddc4b0f4. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 708.116218] env[61995]: DEBUG oslo_concurrency.lockutils [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.120833] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 708.120833] env[61995]: value = "task-794245" [ 708.120833] env[61995]: _type = "Task" [ 708.120833] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.130162] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794245, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.160447] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.160781] env[61995]: DEBUG nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Instance network_info: |[{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 708.161099] env[61995]: DEBUG oslo_concurrency.lockutils [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.161280] env[61995]: DEBUG nova.network.neutron [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Refreshing network info cache for port 802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.162600] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:a7:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ada35c98-01a9-4352-98e4-1d20ba31f928', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '802b752a-238e-4b55-8bd3-f73eddc4b0f4', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.169951] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Creating folder: Project (4c3e03b3825c43ca987d60b339fc0a52). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.172838] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81a5508e-ebec-41cc-9872-dad39a47c067 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.183950] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Created folder: Project (4c3e03b3825c43ca987d60b339fc0a52) in parent group-v185203. [ 708.183950] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Creating folder: Instances. Parent ref: group-v185225. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.183950] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e5c07ab-06c9-442e-86d5-1a9e59f6504b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.192403] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Created folder: Instances in parent group-v185225. [ 708.192629] env[61995]: DEBUG oslo.service.loopingcall [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.192841] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9c53806-68c0-4872-a262-36bc05573674] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 708.192995] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8015c15-293e-4422-91e7-2832058d2ef8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.211658] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.211658] env[61995]: value = "task-794248" [ 708.211658] env[61995]: _type = "Task" [ 708.211658] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.219274] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794248, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.322496] env[61995]: DEBUG nova.compute.utils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.326016] env[61995]: DEBUG nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.326225] env[61995]: DEBUG nova.network.neutron [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 708.461573] env[61995]: DEBUG oslo_vmware.api [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794244, 'name': PowerOffVM_Task, 'duration_secs': 0.167913} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.461961] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 708.463580] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 708.463898] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b2dd7d0-7adc-40e6-b599-02b28e195262 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.466846] env[61995]: DEBUG nova.policy [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19838aa0b8cb4ababef1f656f96fa844', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be010aeecd994d528d00149d329c683f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 708.502180] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 708.502401] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 708.502830] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Deleting the datastore file [datastore1] e78f8718-e0be-4b2a-ab70-9565ce01ab9d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 708.507072] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bca2e607-c703-44fb-afd9-3622e4178bd2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.509693] env[61995]: INFO nova.compute.manager [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Took 33.58 seconds to build instance. [ 708.516656] env[61995]: DEBUG oslo_vmware.api [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for the task: (returnval){ [ 708.516656] env[61995]: value = "task-794250" [ 708.516656] env[61995]: _type = "Task" [ 708.516656] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.530951] env[61995]: DEBUG oslo_vmware.api [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794250, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.630812] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794245, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.681891] env[61995]: DEBUG nova.network.neutron [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updated VIF entry in instance network info cache for port 802b752a-238e-4b55-8bd3-f73eddc4b0f4. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 708.682294] env[61995]: DEBUG nova.network.neutron [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.721279] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794248, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.826687] env[61995]: DEBUG nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.012758] env[61995]: DEBUG oslo_concurrency.lockutils [None req-780b3aed-53b4-4c40-a045-46ff93dbd61d tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.835s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.030763] env[61995]: DEBUG oslo_vmware.api [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Task: {'id': task-794250, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097844} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.031667] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 709.031913] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 709.032129] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 709.032310] env[61995]: INFO nova.compute.manager [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 709.032640] env[61995]: DEBUG oslo.service.loopingcall [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.032971] env[61995]: DEBUG nova.compute.manager [-] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.033088] env[61995]: DEBUG nova.network.neutron [-] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 709.069919] env[61995]: DEBUG nova.network.neutron [-] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.129935] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794245, 'name': ReconfigVM_Task, 'duration_secs': 0.736417} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.130215] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Reconfigured VM instance instance-0000001b to attach disk [datastore2] e53d10a3-1d16-498d-b4d8-abe7205410ec/e53d10a3-1d16-498d-b4d8-abe7205410ec.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 709.132925] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0576fe12-bdc8-4d92-bac4-0adc183b42f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.139914] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 709.139914] env[61995]: value = "task-794251" [ 709.139914] env[61995]: _type = "Task" [ 709.139914] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.148189] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794251, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.185610] env[61995]: DEBUG oslo_concurrency.lockutils [req-f85cbf8d-8f54-4d82-aa4b-c77926f4110f req-380ffad9-5351-4f18-ae03-4956ee3e13ba service nova] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.223622] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794248, 'name': CreateVM_Task, 'duration_secs': 0.741606} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.223778] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9c53806-68c0-4872-a262-36bc05573674] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 709.224761] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.224929] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.225286] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 709.228132] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46b0125c-8aa6-445e-8a74-083a557cd32b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.232628] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 709.232628] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ffa2d8-44c5-56ed-5a54-4558f8815424" [ 709.232628] env[61995]: _type = "Task" [ 709.232628] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.241350] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ffa2d8-44c5-56ed-5a54-4558f8815424, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.257043] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca357bbf-528a-4a25-9a5e-e4d46f30adce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.263789] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293ee8ca-c09c-4766-ae24-0b06bc77fd5e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.293478] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab4eb6b-b4c5-4e65-8479-fc510d0f33a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.301389] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0345e4-2e8e-41fb-a885-2d03b2c5fbad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.315347] env[61995]: DEBUG nova.compute.provider_tree [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.416022] env[61995]: DEBUG nova.network.neutron [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Successfully created port: 1999392b-e938-43fc-9361-7179466a8aea {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.514760] env[61995]: DEBUG nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.572592] env[61995]: DEBUG nova.network.neutron [-] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.650872] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794251, 'name': Rename_Task, 'duration_secs': 0.320471} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.651675] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 709.651918] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ad216ac-cede-48f6-90e4-9a7a0939277e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.658250] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 709.658250] env[61995]: value = "task-794252" [ 709.658250] env[61995]: _type = "Task" [ 709.658250] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.666549] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.743484] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ffa2d8-44c5-56ed-5a54-4558f8815424, 'name': SearchDatastore_Task, 'duration_secs': 0.009027} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.743484] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.743793] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.743936] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.744100] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.744285] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.744560] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0344331-83d3-4aa4-b9e0-1643bb94f328 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.752205] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.752447] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 709.753369] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e79c03b9-349e-4344-a2b2-4ef83a0c7fbc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.759139] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 709.759139] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521999f1-65dd-4da5-7059-cff1b9584aa3" [ 709.759139] env[61995]: _type = "Task" [ 709.759139] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.765899] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521999f1-65dd-4da5-7059-cff1b9584aa3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.820056] env[61995]: DEBUG nova.scheduler.client.report [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.831404] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Successfully updated port: ebbd71fc-00dd-4b27-95ed-6670ac9b552e {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 709.840777] env[61995]: DEBUG nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.868466] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.868733] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.868890] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.869121] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.869286] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.869454] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.869918] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.869918] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.870052] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.870204] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.870369] env[61995]: DEBUG nova.virt.hardware [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.871569] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e90d21-1715-43aa-a45b-767a324bb40f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.879736] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57140d62-6d23-4189-8837-78a17d71979a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.915644] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.915644] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.915644] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.915644] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.915829] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.924647] env[61995]: INFO nova.compute.manager [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Terminating instance [ 709.927941] env[61995]: DEBUG nova.compute.manager [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.928171] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 709.929020] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ff625f-c319-49e5-84a9-b4321087f635 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.938654] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 709.939200] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e7f2a7c-e44c-4914-b294-bef714b2239a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.946944] env[61995]: DEBUG oslo_vmware.api [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 709.946944] env[61995]: value = "task-794253" [ 709.946944] env[61995]: _type = "Task" [ 709.946944] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.958724] env[61995]: DEBUG oslo_vmware.api [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.052849] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.079443] env[61995]: INFO nova.compute.manager [-] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Took 1.04 seconds to deallocate network for instance. [ 710.144266] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.145294] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.145294] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.145413] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.146081] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.150344] env[61995]: INFO nova.compute.manager [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Terminating instance [ 710.159231] env[61995]: DEBUG nova.compute.manager [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 710.159231] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 710.159231] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b71949-7345-431b-8906-f6b5e2a7c29c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.167025] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 710.167025] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de9f2c34-4bb2-45e2-83df-17f1b1a48f8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.171637] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794252, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.176287] env[61995]: DEBUG oslo_vmware.api [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 710.176287] env[61995]: value = "task-794254" [ 710.176287] env[61995]: _type = "Task" [ 710.176287] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.184028] env[61995]: DEBUG oslo_vmware.api [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794254, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.224204] env[61995]: DEBUG nova.compute.manager [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-vif-plugged-ebbd71fc-00dd-4b27-95ed-6670ac9b552e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.224204] env[61995]: DEBUG oslo_concurrency.lockutils [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] Acquiring lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.224204] env[61995]: DEBUG oslo_concurrency.lockutils [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.224204] env[61995]: DEBUG oslo_concurrency.lockutils [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.224204] env[61995]: DEBUG nova.compute.manager [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] No waiting events found dispatching network-vif-plugged-ebbd71fc-00dd-4b27-95ed-6670ac9b552e {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 710.224656] env[61995]: WARNING nova.compute.manager [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received unexpected event network-vif-plugged-ebbd71fc-00dd-4b27-95ed-6670ac9b552e for instance with vm_state building and task_state spawning. [ 710.224656] env[61995]: DEBUG nova.compute.manager [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-changed-ebbd71fc-00dd-4b27-95ed-6670ac9b552e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.224656] env[61995]: DEBUG nova.compute.manager [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Refreshing instance network info cache due to event network-changed-ebbd71fc-00dd-4b27-95ed-6670ac9b552e. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 710.224883] env[61995]: DEBUG oslo_concurrency.lockutils [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] Acquiring lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.225432] env[61995]: DEBUG oslo_concurrency.lockutils [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] Acquired lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.225432] env[61995]: DEBUG nova.network.neutron [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Refreshing network info cache for port ebbd71fc-00dd-4b27-95ed-6670ac9b552e {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.271264] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521999f1-65dd-4da5-7059-cff1b9584aa3, 'name': SearchDatastore_Task, 'duration_secs': 0.009192} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.272585] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2cfdb94-fa99-49bc-a705-427236a98521 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.280026] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 710.280026] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523760e0-89e3-3bd2-6e9b-cd3a11d54e6e" [ 710.280026] env[61995]: _type = "Task" [ 710.280026] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.286572] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523760e0-89e3-3bd2-6e9b-cd3a11d54e6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.324866] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.325707] env[61995]: DEBUG nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.330252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.797s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.331759] env[61995]: INFO nova.compute.claims [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.460667] env[61995]: DEBUG oslo_vmware.api [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794253, 'name': PowerOffVM_Task, 'duration_secs': 0.190042} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.461247] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 710.461525] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 710.461850] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71a7f972-292b-410c-9849-b7878b999801 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.591372] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.635106] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 710.635106] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 710.635106] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleting the datastore file [datastore2] f47e5bbb-f4a8-4686-8195-8348d2ffb08d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 710.635106] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d982f6f-20ad-47a2-ae35-9ea29139613c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.640544] env[61995]: DEBUG oslo_vmware.api [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 710.640544] env[61995]: value = "task-794256" [ 710.640544] env[61995]: _type = "Task" [ 710.640544] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.650023] env[61995]: DEBUG oslo_vmware.api [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794256, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.671018] env[61995]: DEBUG oslo_vmware.api [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794252, 'name': PowerOnVM_Task, 'duration_secs': 0.535259} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.671018] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 710.671018] env[61995]: INFO nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Took 8.48 seconds to spawn the instance on the hypervisor. [ 710.671018] env[61995]: DEBUG nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 710.671018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3ab4e9-6740-4c5a-a2ef-27be17b8e71f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.686329] env[61995]: DEBUG oslo_vmware.api [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794254, 'name': PowerOffVM_Task, 'duration_secs': 0.204673} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.686811] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 710.687150] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 710.687789] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bca8165-9834-46a2-bb14-57ee8e839add {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.763490] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 710.763490] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 710.763805] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleting the datastore file [datastore2] 08489e01-2ce3-45a1-8f87-2a4c89230aa2 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 710.763851] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f0819f6-9bae-47cd-8271-0dd53a7c75c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.770706] env[61995]: DEBUG oslo_vmware.api [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for the task: (returnval){ [ 710.770706] env[61995]: value = "task-794258" [ 710.770706] env[61995]: _type = "Task" [ 710.770706] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.780428] env[61995]: DEBUG oslo_vmware.api [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.792578] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523760e0-89e3-3bd2-6e9b-cd3a11d54e6e, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.792578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.792578] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] d9c53806-68c0-4872-a262-36bc05573674/d9c53806-68c0-4872-a262-36bc05573674.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 710.792578] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08b8d1f5-fb3c-49bf-b073-8b0c67151c0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.799488] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 710.799488] env[61995]: value = "task-794259" [ 710.799488] env[61995]: _type = "Task" [ 710.799488] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.808424] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.809379] env[61995]: DEBUG nova.network.neutron [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.831957] env[61995]: DEBUG nova.compute.utils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.834067] env[61995]: DEBUG nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.834323] env[61995]: DEBUG nova.network.neutron [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 710.895677] env[61995]: DEBUG nova.policy [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2388a3bd1587413285ff9f9ec07739b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9301d14211a1464eb740d9bb745b631a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 711.031798] env[61995]: DEBUG nova.network.neutron [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.152084] env[61995]: DEBUG oslo_vmware.api [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794256, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160289} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.152247] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 711.152548] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 711.152807] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 711.153070] env[61995]: INFO nova.compute.manager [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Took 1.22 seconds to destroy the instance on the hypervisor. [ 711.153330] env[61995]: DEBUG oslo.service.loopingcall [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.153665] env[61995]: DEBUG nova.compute.manager [-] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.154197] env[61995]: DEBUG nova.network.neutron [-] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 711.191109] env[61995]: INFO nova.compute.manager [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Took 32.78 seconds to build instance. [ 711.282640] env[61995]: DEBUG oslo_vmware.api [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Task: {'id': task-794258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155451} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.282640] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 711.284345] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 711.284701] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 711.284988] env[61995]: INFO nova.compute.manager [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 711.285359] env[61995]: DEBUG oslo.service.loopingcall [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.285722] env[61995]: DEBUG nova.compute.manager [-] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.285920] env[61995]: DEBUG nova.network.neutron [-] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 711.310952] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794259, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49005} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.311377] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] d9c53806-68c0-4872-a262-36bc05573674/d9c53806-68c0-4872-a262-36bc05573674.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 711.311714] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 711.312057] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4727ee7e-bf26-4de0-ac86-14f57b0113e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.318621] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 711.318621] env[61995]: value = "task-794260" [ 711.318621] env[61995]: _type = "Task" [ 711.318621] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.328330] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794260, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.329355] env[61995]: DEBUG nova.network.neutron [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Successfully created port: 515eb152-ff0c-4c68-896f-d46c12f8c848 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.337194] env[61995]: DEBUG nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.541244] env[61995]: DEBUG oslo_concurrency.lockutils [req-97dd0158-fbe6-4fb6-a0bd-1ab4d6fde8da req-30ac10cf-f6d7-413a-b76e-18545ac11577 service nova] Releasing lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.693542] env[61995]: DEBUG oslo_concurrency.lockutils [None req-078b2263-a892-4d92-a749-0dd240c15eee tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.712s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.750135] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48649de-0d82-4a64-aeff-201b2d947a7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.757712] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd0fc3a-af3a-4a72-8174-261832b7c968 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.789897] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8f12fb-6de8-458e-a9d9-98bbbb609857 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.798729] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4143efe-a606-46cf-a08a-e5ccb1862828 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.812938] env[61995]: DEBUG nova.compute.provider_tree [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.827928] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794260, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06949} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.828226] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 711.829012] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d7f544-728f-4515-9e7b-831b7607a418 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.851754] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] d9c53806-68c0-4872-a262-36bc05573674/d9c53806-68c0-4872-a262-36bc05573674.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 711.855055] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b15f722f-6eab-486c-8d5f-66893ce54470 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.874773] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 711.874773] env[61995]: value = "task-794261" [ 711.874773] env[61995]: _type = "Task" [ 711.874773] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.883029] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794261, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.199549] env[61995]: DEBUG nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 712.304257] env[61995]: DEBUG nova.network.neutron [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Successfully updated port: 1999392b-e938-43fc-9361-7179466a8aea {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 712.316965] env[61995]: DEBUG nova.scheduler.client.report [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.371040] env[61995]: DEBUG nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.384942] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794261, 'name': ReconfigVM_Task, 'duration_secs': 0.307954} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.385252] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Reconfigured VM instance instance-0000001c to attach disk [datastore2] d9c53806-68c0-4872-a262-36bc05573674/d9c53806-68c0-4872-a262-36bc05573674.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.385851] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57c5d96e-7b1e-463a-93c0-cc6c1c507522 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.392922] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 712.392922] env[61995]: value = "task-794262" [ 712.392922] env[61995]: _type = "Task" [ 712.392922] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.400717] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.400939] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.401108] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.401291] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.401434] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.401576] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.401774] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.401926] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.402105] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.402265] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.402428] env[61995]: DEBUG nova.virt.hardware [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.403634] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f87282-3ca1-4fd6-81ca-f9b5c9dcbed5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.412165] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794262, 'name': Rename_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.416551] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a8644a-1241-4275-be78-8bd9aa4012ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.584619] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Successfully updated port: da35a09f-ea6a-43b8-a817-8a2efbe499bb {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 712.661833] env[61995]: DEBUG nova.network.neutron [-] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.668198] env[61995]: DEBUG nova.network.neutron [-] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.736627] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.748485] env[61995]: DEBUG nova.compute.manager [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Received event network-vif-plugged-1999392b-e938-43fc-9361-7179466a8aea {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.748667] env[61995]: DEBUG oslo_concurrency.lockutils [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] Acquiring lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.748848] env[61995]: DEBUG oslo_concurrency.lockutils [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.749023] env[61995]: DEBUG oslo_concurrency.lockutils [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.749189] env[61995]: DEBUG nova.compute.manager [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] No waiting events found dispatching network-vif-plugged-1999392b-e938-43fc-9361-7179466a8aea {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 712.749459] env[61995]: WARNING nova.compute.manager [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Received unexpected event network-vif-plugged-1999392b-e938-43fc-9361-7179466a8aea for instance with vm_state building and task_state spawning. [ 712.749538] env[61995]: DEBUG nova.compute.manager [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Received event network-vif-deleted-7e64a241-d943-4260-9058-75a25d87a679 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.749661] env[61995]: DEBUG nova.compute.manager [req-27f7473d-e050-4184-bd90-ada0627d1aa5 req-9faefec6-4a04-4db9-8faf-d895bef6d827 service nova] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Received event network-vif-deleted-71f99164-4e02-4b29-bb4c-7a576848a8ee {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.807542] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "refresh_cache-24db5fad-52fd-4689-a8aa-2ae4b0a06710" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.807542] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquired lock "refresh_cache-24db5fad-52fd-4689-a8aa-2ae4b0a06710" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.807542] env[61995]: DEBUG nova.network.neutron [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.825619] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.826190] env[61995]: DEBUG nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.830204] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.597s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.832752] env[61995]: INFO nova.compute.claims [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.904276] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794262, 'name': Rename_Task, 'duration_secs': 0.151289} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.904539] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 712.904777] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14c70ece-6c4b-4661-8ea8-a6e180718bdc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.911307] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 712.911307] env[61995]: value = "task-794263" [ 712.911307] env[61995]: _type = "Task" [ 712.911307] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.919149] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794263, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.166705] env[61995]: INFO nova.compute.manager [-] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Took 2.01 seconds to deallocate network for instance. [ 713.172021] env[61995]: INFO nova.compute.manager [-] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Took 1.89 seconds to deallocate network for instance. [ 713.328936] env[61995]: DEBUG nova.network.neutron [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Successfully updated port: 515eb152-ff0c-4c68-896f-d46c12f8c848 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 713.332558] env[61995]: DEBUG nova.compute.utils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.335652] env[61995]: DEBUG nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.335652] env[61995]: DEBUG nova.network.neutron [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.394581] env[61995]: DEBUG nova.policy [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 713.407971] env[61995]: DEBUG nova.network.neutron [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.423973] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794263, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.652212] env[61995]: DEBUG nova.network.neutron [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Updating instance_info_cache with network_info: [{"id": "1999392b-e938-43fc-9361-7179466a8aea", "address": "fa:16:3e:ea:b7:02", "network": {"id": "8e7a3e75-8010-41df-ba79-bc8e78767dc9", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-625447152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be010aeecd994d528d00149d329c683f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1999392b-e9", "ovs_interfaceid": "1999392b-e938-43fc-9361-7179466a8aea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.673608] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.682095] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.700759] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "d056dadd-492d-4a4d-abc2-fc5f113f865b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.701330] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.759146] env[61995]: DEBUG nova.compute.manager [req-73e3097b-4078-4e62-a5b6-11cd05d7957f req-f6bc3f3d-8460-4717-87d5-92bd4ff09360 service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Received event network-vif-plugged-515eb152-ff0c-4c68-896f-d46c12f8c848 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 713.759146] env[61995]: DEBUG oslo_concurrency.lockutils [req-73e3097b-4078-4e62-a5b6-11cd05d7957f req-f6bc3f3d-8460-4717-87d5-92bd4ff09360 service nova] Acquiring lock "45fdae21-bec2-438a-8f2a-c23b0e428057-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.759146] env[61995]: DEBUG oslo_concurrency.lockutils [req-73e3097b-4078-4e62-a5b6-11cd05d7957f req-f6bc3f3d-8460-4717-87d5-92bd4ff09360 service nova] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.759146] env[61995]: DEBUG oslo_concurrency.lockutils [req-73e3097b-4078-4e62-a5b6-11cd05d7957f req-f6bc3f3d-8460-4717-87d5-92bd4ff09360 service nova] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.759146] env[61995]: DEBUG nova.compute.manager [req-73e3097b-4078-4e62-a5b6-11cd05d7957f req-f6bc3f3d-8460-4717-87d5-92bd4ff09360 service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] No waiting events found dispatching network-vif-plugged-515eb152-ff0c-4c68-896f-d46c12f8c848 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 713.760093] env[61995]: WARNING nova.compute.manager [req-73e3097b-4078-4e62-a5b6-11cd05d7957f req-f6bc3f3d-8460-4717-87d5-92bd4ff09360 service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Received unexpected event network-vif-plugged-515eb152-ff0c-4c68-896f-d46c12f8c848 for instance with vm_state building and task_state spawning. [ 713.817189] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.817189] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.837020] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-45fdae21-bec2-438a-8f2a-c23b0e428057" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.837181] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-45fdae21-bec2-438a-8f2a-c23b0e428057" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.837335] env[61995]: DEBUG nova.network.neutron [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.844062] env[61995]: DEBUG nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.875270] env[61995]: DEBUG nova.network.neutron [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Successfully created port: 0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.925484] env[61995]: DEBUG oslo_vmware.api [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794263, 'name': PowerOnVM_Task, 'duration_secs': 0.889884} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.925779] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 713.926105] env[61995]: INFO nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Took 9.18 seconds to spawn the instance on the hypervisor. [ 713.926282] env[61995]: DEBUG nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 713.927273] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5538f48c-7bb2-4993-9a80-8cd078ca647f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.155432] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Releasing lock "refresh_cache-24db5fad-52fd-4689-a8aa-2ae4b0a06710" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.155432] env[61995]: DEBUG nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Instance network_info: |[{"id": "1999392b-e938-43fc-9361-7179466a8aea", "address": "fa:16:3e:ea:b7:02", "network": {"id": "8e7a3e75-8010-41df-ba79-bc8e78767dc9", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-625447152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be010aeecd994d528d00149d329c683f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1999392b-e9", "ovs_interfaceid": "1999392b-e938-43fc-9361-7179466a8aea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 714.155821] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:b7:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b931c4c-f73c-4fbd-9c9f-0270834cc69e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1999392b-e938-43fc-9361-7179466a8aea', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.163409] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Creating folder: Project (be010aeecd994d528d00149d329c683f). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 714.165900] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc4228dd-c7bd-49c8-87ce-f1081f750557 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.176428] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Created folder: Project (be010aeecd994d528d00149d329c683f) in parent group-v185203. [ 714.176604] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Creating folder: Instances. Parent ref: group-v185228. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 714.176832] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83049e00-08cd-42d4-b4e1-7ea34c116edf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.188314] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Created folder: Instances in parent group-v185228. [ 714.188543] env[61995]: DEBUG oslo.service.loopingcall [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.190764] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 714.191764] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f1f925e-a055-42bc-9aa5-4b38055617d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.215260] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.215260] env[61995]: value = "task-794266" [ 714.215260] env[61995]: _type = "Task" [ 714.215260] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.223412] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794266, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.263212] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a246ac40-c85c-4fd5-9392-871a103d6130 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.270577] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7de83c-1931-4c41-b557-4107ca57280f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.305342] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82fce3bc-c076-442f-9533-d3c7e9604a79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.313292] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8826588a-fe15-4258-8a86-f0cc907ba9d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.326809] env[61995]: DEBUG nova.compute.provider_tree [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.380154] env[61995]: DEBUG nova.network.neutron [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.449825] env[61995]: INFO nova.compute.manager [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Took 33.98 seconds to build instance. [ 714.543008] env[61995]: DEBUG nova.network.neutron [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Updating instance_info_cache with network_info: [{"id": "515eb152-ff0c-4c68-896f-d46c12f8c848", "address": "fa:16:3e:14:7f:47", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap515eb152-ff", "ovs_interfaceid": "515eb152-ff0c-4c68-896f-d46c12f8c848", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.726443] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794266, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.782927] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Received event network-changed-1999392b-e938-43fc-9361-7179466a8aea {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.783297] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Refreshing instance network info cache due to event network-changed-1999392b-e938-43fc-9361-7179466a8aea. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 714.783588] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Acquiring lock "refresh_cache-24db5fad-52fd-4689-a8aa-2ae4b0a06710" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.783811] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Acquired lock "refresh_cache-24db5fad-52fd-4689-a8aa-2ae4b0a06710" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.784119] env[61995]: DEBUG nova.network.neutron [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Refreshing network info cache for port 1999392b-e938-43fc-9361-7179466a8aea {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.832069] env[61995]: DEBUG nova.scheduler.client.report [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.858963] env[61995]: DEBUG nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.883792] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.883970] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.884214] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.884330] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.884455] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.884643] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.884797] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.884987] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.885138] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.885304] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.885501] env[61995]: DEBUG nova.virt.hardware [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.886454] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50045d15-5c5f-4117-848e-826a185697ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.889785] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Successfully updated port: b569b424-48e7-45f4-8f0d-53192e53abf0 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 714.896544] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2439323-9dc5-4d1d-9003-45d041e8c423 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.951252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-777c2963-941e-4387-9383-5a85afdb1f19 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.863s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.049412] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-45fdae21-bec2-438a-8f2a-c23b0e428057" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.049776] env[61995]: DEBUG nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Instance network_info: |[{"id": "515eb152-ff0c-4c68-896f-d46c12f8c848", "address": "fa:16:3e:14:7f:47", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap515eb152-ff", "ovs_interfaceid": "515eb152-ff0c-4c68-896f-d46c12f8c848", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 715.050403] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:7f:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '515eb152-ff0c-4c68-896f-d46c12f8c848', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 715.057445] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating folder: Project (9301d14211a1464eb740d9bb745b631a). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 715.057714] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46632b77-6e99-466d-b1bb-2411db6ef058 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.068570] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created folder: Project (9301d14211a1464eb740d9bb745b631a) in parent group-v185203. [ 715.068750] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating folder: Instances. Parent ref: group-v185231. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 715.068999] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0e07736-de97-4cd9-bf6b-8375974ca365 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.077390] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created folder: Instances in parent group-v185231. [ 715.077618] env[61995]: DEBUG oslo.service.loopingcall [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.077812] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 715.078060] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d91eb19-eb28-4dc6-88da-dcb942f746f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.097603] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 715.097603] env[61995]: value = "task-794269" [ 715.097603] env[61995]: _type = "Task" [ 715.097603] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.105461] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794269, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.226455] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794266, 'name': CreateVM_Task, 'duration_secs': 0.541641} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.226703] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 715.227477] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.227710] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.228154] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 715.228462] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38209cab-d673-425f-8204-c60b16aecef5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.233738] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 715.233738] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f52672-111b-68f1-3e4b-d4ffd3d08d66" [ 715.233738] env[61995]: _type = "Task" [ 715.233738] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.244564] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f52672-111b-68f1-3e4b-d4ffd3d08d66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.338527] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.339396] env[61995]: DEBUG nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.342013] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.423s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.346315] env[61995]: DEBUG nova.objects.instance [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lazy-loading 'resources' on Instance uuid 06496d6a-324d-41d4-b3e5-52e3d93636bb {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 715.402719] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.403122] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.403507] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.456820] env[61995]: DEBUG nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.589498] env[61995]: DEBUG nova.network.neutron [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Updated VIF entry in instance network info cache for port 1999392b-e938-43fc-9361-7179466a8aea. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 715.590072] env[61995]: DEBUG nova.network.neutron [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Updating instance_info_cache with network_info: [{"id": "1999392b-e938-43fc-9361-7179466a8aea", "address": "fa:16:3e:ea:b7:02", "network": {"id": "8e7a3e75-8010-41df-ba79-bc8e78767dc9", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-625447152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be010aeecd994d528d00149d329c683f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1999392b-e9", "ovs_interfaceid": "1999392b-e938-43fc-9361-7179466a8aea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.612980] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794269, 'name': CreateVM_Task, 'duration_secs': 0.289091} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.612980] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 715.612980] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.745531] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f52672-111b-68f1-3e4b-d4ffd3d08d66, 'name': SearchDatastore_Task, 'duration_secs': 0.012747} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.746074] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.746311] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.746537] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.746676] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.747061] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 715.747343] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.747633] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 715.747845] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b5268c5-24f2-48af-88eb-e0e70bfed08b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.754376] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4461dba9-af25-42ff-9ea1-411e09ae1e66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.757553] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 715.757553] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e0e97c-b906-9ff2-b427-9596873faeb4" [ 715.757553] env[61995]: _type = "Task" [ 715.757553] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.767460] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.767460] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 715.768370] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe058eb1-9910-4459-8e07-69b6683b9f6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.775245] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e0e97c-b906-9ff2-b427-9596873faeb4, 'name': SearchDatastore_Task, 'duration_secs': 0.009074} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.776105] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.776391] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.776628] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.777987] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 715.777987] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525b0e6d-babc-0fde-eedb-e1f6f694ca94" [ 715.777987] env[61995]: _type = "Task" [ 715.777987] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.784859] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525b0e6d-babc-0fde-eedb-e1f6f694ca94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.832568] env[61995]: DEBUG nova.compute.manager [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Received event network-changed-515eb152-ff0c-4c68-896f-d46c12f8c848 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 715.832811] env[61995]: DEBUG nova.compute.manager [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Refreshing instance network info cache due to event network-changed-515eb152-ff0c-4c68-896f-d46c12f8c848. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 715.832968] env[61995]: DEBUG oslo_concurrency.lockutils [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] Acquiring lock "refresh_cache-45fdae21-bec2-438a-8f2a-c23b0e428057" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.833327] env[61995]: DEBUG oslo_concurrency.lockutils [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] Acquired lock "refresh_cache-45fdae21-bec2-438a-8f2a-c23b0e428057" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.833327] env[61995]: DEBUG nova.network.neutron [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Refreshing network info cache for port 515eb152-ff0c-4c68-896f-d46c12f8c848 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 715.849430] env[61995]: DEBUG nova.compute.utils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.850672] env[61995]: DEBUG nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.850836] env[61995]: DEBUG nova.network.neutron [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.910344] env[61995]: DEBUG nova.policy [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18d2cb50dbc24b87983ac8f24b0ebe09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '64e0bdabc3cf4eb7b606adda78f9a0b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 715.979473] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.991098] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.002809] env[61995]: DEBUG nova.network.neutron [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Successfully updated port: 0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 716.095033] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Releasing lock "refresh_cache-24db5fad-52fd-4689-a8aa-2ae4b0a06710" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.095033] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-vif-plugged-da35a09f-ea6a-43b8-a817-8a2efbe499bb {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.095033] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Acquiring lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.095033] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.095033] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.095779] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] No waiting events found dispatching network-vif-plugged-da35a09f-ea6a-43b8-a817-8a2efbe499bb {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 716.095779] env[61995]: WARNING nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received unexpected event network-vif-plugged-da35a09f-ea6a-43b8-a817-8a2efbe499bb for instance with vm_state building and task_state spawning. [ 716.095779] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-changed-da35a09f-ea6a-43b8-a817-8a2efbe499bb {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.095779] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Refreshing instance network info cache due to event network-changed-da35a09f-ea6a-43b8-a817-8a2efbe499bb. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 716.095779] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Acquiring lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.289319] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525b0e6d-babc-0fde-eedb-e1f6f694ca94, 'name': SearchDatastore_Task, 'duration_secs': 0.007845} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.290123] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00d40918-61dc-418c-87e9-0b59b38de1d9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.295375] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 716.295375] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520e4681-1c0f-4185-e315-3633374dfba1" [ 716.295375] env[61995]: _type = "Task" [ 716.295375] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.303479] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520e4681-1c0f-4185-e315-3633374dfba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.357786] env[61995]: DEBUG nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.392820] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdf64cb-d11a-4855-9285-7908f1b01edd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.402865] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ac67bc-99a2-4729-93ff-510da1ce552d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.433790] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b4b3fe-273b-42db-87ae-e0645993a9bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.443182] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a166b543-9f4c-469d-b114-0d5af6c46cb4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.459289] env[61995]: DEBUG nova.compute.provider_tree [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.509540] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.509803] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.509996] env[61995]: DEBUG nova.network.neutron [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.673488] env[61995]: DEBUG nova.network.neutron [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Updated VIF entry in instance network info cache for port 515eb152-ff0c-4c68-896f-d46c12f8c848. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 716.673940] env[61995]: DEBUG nova.network.neutron [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Updating instance_info_cache with network_info: [{"id": "515eb152-ff0c-4c68-896f-d46c12f8c848", "address": "fa:16:3e:14:7f:47", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap515eb152-ff", "ovs_interfaceid": "515eb152-ff0c-4c68-896f-d46c12f8c848", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.746482] env[61995]: DEBUG nova.network.neutron [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Successfully created port: f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.809041] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520e4681-1c0f-4185-e315-3633374dfba1, 'name': SearchDatastore_Task, 'duration_secs': 0.00975} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.809041] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.809041] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 24db5fad-52fd-4689-a8aa-2ae4b0a06710/24db5fad-52fd-4689-a8aa-2ae4b0a06710.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 716.809041] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.809261] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.809261] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbb6336e-cd3e-4bad-8f97-1c262edda8cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.809921] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-717bc0db-423e-4018-bdb6-d867b63fe75d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.820570] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.820826] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 716.821641] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 716.821641] env[61995]: value = "task-794270" [ 716.821641] env[61995]: _type = "Task" [ 716.821641] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.821991] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca835456-2783-4034-aebe-bf465c28a34c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.830416] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 716.830416] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52802a8d-d9c7-b78c-49f1-d03b9e664103" [ 716.830416] env[61995]: _type = "Task" [ 716.830416] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.834949] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.843347] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52802a8d-d9c7-b78c-49f1-d03b9e664103, 'name': SearchDatastore_Task, 'duration_secs': 0.008429} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.844168] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b29efe36-9ee1-4ae9-93b5-e7cb561f6e0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.850024] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 716.850024] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b4eaee-9ac8-07f5-cf00-76fa3f93f4e8" [ 716.850024] env[61995]: _type = "Task" [ 716.850024] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.861540] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b4eaee-9ac8-07f5-cf00-76fa3f93f4e8, 'name': SearchDatastore_Task, 'duration_secs': 0.007997} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.861832] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.862121] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 716.862434] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd3dc4ac-f819-4a33-abf8-92d2253338f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.875348] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 716.875348] env[61995]: value = "task-794271" [ 716.875348] env[61995]: _type = "Task" [ 716.875348] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.884794] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794271, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.963782] env[61995]: DEBUG nova.scheduler.client.report [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.969789] env[61995]: DEBUG nova.compute.manager [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-changed-b569b424-48e7-45f4-8f0d-53192e53abf0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.969997] env[61995]: DEBUG nova.compute.manager [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Refreshing instance network info cache due to event network-changed-b569b424-48e7-45f4-8f0d-53192e53abf0. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 716.970399] env[61995]: DEBUG oslo_concurrency.lockutils [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] Acquiring lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.065961] env[61995]: DEBUG nova.network.neutron [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.121594] env[61995]: DEBUG nova.network.neutron [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updating instance_info_cache with network_info: [{"id": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "address": "fa:16:3e:96:46:60", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbd71fc-00", "ovs_interfaceid": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "address": "fa:16:3e:f5:db:69", "network": {"id": "3a407f9b-a9f0-4839-8dd2-f10ecd904b43", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1652315845", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda35a09f-ea", "ovs_interfaceid": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b569b424-48e7-45f4-8f0d-53192e53abf0", "address": "fa:16:3e:53:7c:52", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.192", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb569b424-48", "ovs_interfaceid": "b569b424-48e7-45f4-8f0d-53192e53abf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.180742] env[61995]: DEBUG oslo_concurrency.lockutils [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] Releasing lock "refresh_cache-45fdae21-bec2-438a-8f2a-c23b0e428057" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.181038] env[61995]: DEBUG nova.compute.manager [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received event network-vif-plugged-0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.181247] env[61995]: DEBUG oslo_concurrency.lockutils [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.181455] env[61995]: DEBUG oslo_concurrency.lockutils [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.181617] env[61995]: DEBUG oslo_concurrency.lockutils [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.181778] env[61995]: DEBUG nova.compute.manager [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] No waiting events found dispatching network-vif-plugged-0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 717.181946] env[61995]: WARNING nova.compute.manager [req-8dec4f5d-1a68-4636-af86-ef181cc094a3 req-7856f54f-488f-4481-8137-ac769173300e service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received unexpected event network-vif-plugged-0f5a4362-281d-470a-95b7-6214553bfc8a for instance with vm_state building and task_state spawning. [ 717.266874] env[61995]: DEBUG nova.network.neutron [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.336091] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794270, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504898} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.336410] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 24db5fad-52fd-4689-a8aa-2ae4b0a06710/24db5fad-52fd-4689-a8aa-2ae4b0a06710.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 717.336622] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.336893] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5c0ce00-f3f8-439a-9bd4-fac9fdae0bc1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.345069] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 717.345069] env[61995]: value = "task-794272" [ 717.345069] env[61995]: _type = "Task" [ 717.345069] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.356245] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794272, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.371714] env[61995]: DEBUG nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.386737] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794271, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.402637] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.402793] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.403622] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.403622] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.403622] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.403622] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.403622] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.403874] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.403949] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.404182] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.404387] env[61995]: DEBUG nova.virt.hardware [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.405231] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9600810c-6660-46fd-92e8-b8a0608eab28 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.415232] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af632dc3-1eea-4598-8ce9-5e3137a5d03e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.474316] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.132s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.477040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.084s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.479441] env[61995]: INFO nova.compute.claims [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.501171] env[61995]: INFO nova.scheduler.client.report [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Deleted allocations for instance 06496d6a-324d-41d4-b3e5-52e3d93636bb [ 717.623932] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Releasing lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.624367] env[61995]: DEBUG nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Instance network_info: |[{"id": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "address": "fa:16:3e:96:46:60", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbd71fc-00", "ovs_interfaceid": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "address": "fa:16:3e:f5:db:69", "network": {"id": "3a407f9b-a9f0-4839-8dd2-f10ecd904b43", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1652315845", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda35a09f-ea", "ovs_interfaceid": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b569b424-48e7-45f4-8f0d-53192e53abf0", "address": "fa:16:3e:53:7c:52", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.192", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb569b424-48", "ovs_interfaceid": "b569b424-48e7-45f4-8f0d-53192e53abf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 717.624666] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Acquired lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.624844] env[61995]: DEBUG nova.network.neutron [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Refreshing network info cache for port da35a09f-ea6a-43b8-a817-8a2efbe499bb {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 717.626079] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:46:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebbd71fc-00dd-4b27-95ed-6670ac9b552e', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:db:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da35a09f-ea6a-43b8-a817-8a2efbe499bb', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:7c:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b569b424-48e7-45f4-8f0d-53192e53abf0', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.636627] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Creating folder: Project (f22192b1bd12430e9f2de0288d1e7e2a). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.640293] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44891830-bac9-4488-b590-d832ba666f44 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.650407] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Created folder: Project (f22192b1bd12430e9f2de0288d1e7e2a) in parent group-v185203. [ 717.650619] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Creating folder: Instances. Parent ref: group-v185234. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.650875] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08f99c70-d55f-4bec-9adb-735e330c2344 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.660896] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Created folder: Instances in parent group-v185234. [ 717.660896] env[61995]: DEBUG oslo.service.loopingcall [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.660896] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 717.660896] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b212245-b23c-4028-a316-e25be98191dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.689669] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.689669] env[61995]: value = "task-794275" [ 717.689669] env[61995]: _type = "Task" [ 717.689669] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.698476] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794275, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.770740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.771098] env[61995]: DEBUG nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Instance network_info: |[{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 717.771589] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:ac:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f5a4362-281d-470a-95b7-6214553bfc8a', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.779592] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Creating folder: Project (a660ec6d4d7e4e76827642cf247f53c9). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.782291] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81bd3c93-8da1-4200-9370-79c76dff3753 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.793582] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Created folder: Project (a660ec6d4d7e4e76827642cf247f53c9) in parent group-v185203. [ 717.793781] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Creating folder: Instances. Parent ref: group-v185237. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.794149] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11359e8b-b2b4-4d66-91b2-1895afe2c059 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.804124] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Created folder: Instances in parent group-v185237. [ 717.804124] env[61995]: DEBUG oslo.service.loopingcall [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.804124] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 717.804124] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1117132-2817-4e6a-ab91-1b61b507afd3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.824701] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.824701] env[61995]: value = "task-794278" [ 717.824701] env[61995]: _type = "Task" [ 717.824701] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.832385] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794278, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.856026] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794272, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083694} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.856402] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 717.857069] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f1814c-19c3-4f6c-ad29-d0b060875ba4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.879969] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 24db5fad-52fd-4689-a8aa-2ae4b0a06710/24db5fad-52fd-4689-a8aa-2ae4b0a06710.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 717.884102] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42b41371-7805-408d-8733-efb668611604 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.899893] env[61995]: DEBUG nova.compute.manager [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received event network-changed-0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.902215] env[61995]: DEBUG nova.compute.manager [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Refreshing instance network info cache due to event network-changed-0f5a4362-281d-470a-95b7-6214553bfc8a. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 717.902417] env[61995]: DEBUG oslo_concurrency.lockutils [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] Acquiring lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.902561] env[61995]: DEBUG oslo_concurrency.lockutils [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] Acquired lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.902720] env[61995]: DEBUG nova.network.neutron [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Refreshing network info cache for port 0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 717.915423] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794271, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691605} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.917032] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 717.917191] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.917454] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 717.917454] env[61995]: value = "task-794279" [ 717.917454] env[61995]: _type = "Task" [ 717.917454] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.917865] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07f75e68-3159-46e2-a003-c55e70733398 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.928565] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794279, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.929886] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 717.929886] env[61995]: value = "task-794280" [ 717.929886] env[61995]: _type = "Task" [ 717.929886] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.939171] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794280, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.013591] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bccd21b4-4146-4412-8c5d-496c2d9f3977 tempest-DeleteServersAdminTestJSON-815584122 tempest-DeleteServersAdminTestJSON-815584122-project-member] Lock "06496d6a-324d-41d4-b3e5-52e3d93636bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.315s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.044481] env[61995]: DEBUG nova.network.neutron [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updated VIF entry in instance network info cache for port da35a09f-ea6a-43b8-a817-8a2efbe499bb. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 718.044481] env[61995]: DEBUG nova.network.neutron [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updating instance_info_cache with network_info: [{"id": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "address": "fa:16:3e:96:46:60", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbd71fc-00", "ovs_interfaceid": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "address": "fa:16:3e:f5:db:69", "network": {"id": "3a407f9b-a9f0-4839-8dd2-f10ecd904b43", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1652315845", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda35a09f-ea", "ovs_interfaceid": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b569b424-48e7-45f4-8f0d-53192e53abf0", "address": "fa:16:3e:53:7c:52", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.192", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb569b424-48", "ovs_interfaceid": "b569b424-48e7-45f4-8f0d-53192e53abf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.200599] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794275, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.334203] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794278, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.431106] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794279, 'name': ReconfigVM_Task, 'duration_secs': 0.49791} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.431106] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 24db5fad-52fd-4689-a8aa-2ae4b0a06710/24db5fad-52fd-4689-a8aa-2ae4b0a06710.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.434013] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68f79fc0-f9ec-4b79-a861-7b3500fdf8a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.441196] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794280, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.152169} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.441730] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.442044] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 718.442044] env[61995]: value = "task-794281" [ 718.442044] env[61995]: _type = "Task" [ 718.442044] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.442688] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b803e71c-56b7-44b8-8dde-c59e652bd081 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.467420] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.473429] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76c41b21-d343-4770-81e0-694020da61b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.487603] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794281, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.495928] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 718.495928] env[61995]: value = "task-794282" [ 718.495928] env[61995]: _type = "Task" [ 718.495928] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.504546] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794282, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.548158] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Releasing lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.548158] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-vif-plugged-b569b424-48e7-45f4-8f0d-53192e53abf0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 718.548158] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Acquiring lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.548158] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.548817] env[61995]: DEBUG oslo_concurrency.lockutils [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.549147] env[61995]: DEBUG nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] No waiting events found dispatching network-vif-plugged-b569b424-48e7-45f4-8f0d-53192e53abf0 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 718.549485] env[61995]: WARNING nova.compute.manager [req-ec961d0f-990c-4cf7-b2db-146930a6af7c req-bf3ad121-f018-4258-84eb-b4b283cb5e9e service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received unexpected event network-vif-plugged-b569b424-48e7-45f4-8f0d-53192e53abf0 for instance with vm_state building and task_state spawning. [ 718.550170] env[61995]: DEBUG oslo_concurrency.lockutils [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] Acquired lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.550475] env[61995]: DEBUG nova.network.neutron [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Refreshing network info cache for port b569b424-48e7-45f4-8f0d-53192e53abf0 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.669116] env[61995]: DEBUG nova.network.neutron [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Successfully updated port: f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 718.700872] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794275, 'name': CreateVM_Task, 'duration_secs': 0.742673} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.704093] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 718.705076] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.705259] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.705603] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 718.705872] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2c20e9d-c579-47af-b248-22435b3a0340 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.710903] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 718.710903] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c9dbb7-affd-abdc-8008-8de76de1d20a" [ 718.710903] env[61995]: _type = "Task" [ 718.710903] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.721304] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c9dbb7-affd-abdc-8008-8de76de1d20a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.800481] env[61995]: DEBUG nova.network.neutron [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updated VIF entry in instance network info cache for port 0f5a4362-281d-470a-95b7-6214553bfc8a. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 718.800888] env[61995]: DEBUG nova.network.neutron [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.836183] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794278, 'name': CreateVM_Task, 'duration_secs': 0.574256} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.838385] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 718.839290] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.839386] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.839696] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 718.839944] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b694248-6f58-451e-9a39-5464faf36fc1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.844617] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 718.844617] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52294508-326b-679e-a88a-56fb92b61b8a" [ 718.844617] env[61995]: _type = "Task" [ 718.844617] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.852434] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52294508-326b-679e-a88a-56fb92b61b8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.903080] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99820400-32ea-4a48-b14f-08f571063951 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.910327] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9847d0-3d1b-4a2e-924c-0396d7b44a62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.946292] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30360190-76ca-4b8c-ba7f-b2ed1ffa1301 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.960299] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6110eb27-9e7d-409b-8008-6b9838e7b0b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.964100] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794281, 'name': Rename_Task, 'duration_secs': 0.14143} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.964375] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 718.964950] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f096fca-f951-4965-97b3-da0433c781f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.976861] env[61995]: DEBUG nova.compute.provider_tree [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.983572] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 718.983572] env[61995]: value = "task-794283" [ 718.983572] env[61995]: _type = "Task" [ 718.983572] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.992489] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.004811] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794282, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.172658] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "refresh_cache-19fd0358-32e4-434f-b0b3-c8cc6c07ba51" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.172658] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquired lock "refresh_cache-19fd0358-32e4-434f-b0b3-c8cc6c07ba51" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.172658] env[61995]: DEBUG nova.network.neutron [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.224974] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c9dbb7-affd-abdc-8008-8de76de1d20a, 'name': SearchDatastore_Task, 'duration_secs': 0.008926} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.228562] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.229123] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.229584] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.229863] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.230198] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.230677] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c2499ae-2956-47b6-8525-772d0000030f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.236492] env[61995]: DEBUG nova.compute.manager [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Received event network-vif-plugged-f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.236886] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] Acquiring lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.238809] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.238809] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.238809] env[61995]: DEBUG nova.compute.manager [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] No waiting events found dispatching network-vif-plugged-f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 719.238809] env[61995]: WARNING nova.compute.manager [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Received unexpected event network-vif-plugged-f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 for instance with vm_state building and task_state spawning. [ 719.238809] env[61995]: DEBUG nova.compute.manager [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Received event network-changed-f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.238809] env[61995]: DEBUG nova.compute.manager [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Refreshing instance network info cache due to event network-changed-f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 719.238809] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] Acquiring lock "refresh_cache-19fd0358-32e4-434f-b0b3-c8cc6c07ba51" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.247037] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.247037] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 719.247037] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-447f5f8d-b05f-4be9-9a6a-c5434fee6ff2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.252309] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 719.252309] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e05c68-f07f-bdca-644f-36ef7e63dcfb" [ 719.252309] env[61995]: _type = "Task" [ 719.252309] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.261033] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e05c68-f07f-bdca-644f-36ef7e63dcfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.304513] env[61995]: DEBUG oslo_concurrency.lockutils [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] Releasing lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.304944] env[61995]: DEBUG nova.compute.manager [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Received event network-changed-802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.306667] env[61995]: DEBUG nova.compute.manager [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Refreshing instance network info cache due to event network-changed-802b752a-238e-4b55-8bd3-f73eddc4b0f4. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 719.306667] env[61995]: DEBUG oslo_concurrency.lockutils [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.306667] env[61995]: DEBUG oslo_concurrency.lockutils [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.306667] env[61995]: DEBUG nova.network.neutron [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Refreshing network info cache for port 802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 719.356078] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52294508-326b-679e-a88a-56fb92b61b8a, 'name': SearchDatastore_Task, 'duration_secs': 0.008938} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.357847] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.357847] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.357847] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.357847] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.357847] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.357847] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-476f80de-9b5a-46f8-927c-a2ef05264619 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.365631] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.365810] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 719.366517] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52ce87a9-3ba5-446f-9e3c-c2ceb518e5e6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.372523] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 719.372523] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5229c3ac-09b8-1aa7-e488-ccd776620b10" [ 719.372523] env[61995]: _type = "Task" [ 719.372523] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.379761] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5229c3ac-09b8-1aa7-e488-ccd776620b10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.385855] env[61995]: DEBUG nova.network.neutron [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updated VIF entry in instance network info cache for port b569b424-48e7-45f4-8f0d-53192e53abf0. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 719.386355] env[61995]: DEBUG nova.network.neutron [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updating instance_info_cache with network_info: [{"id": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "address": "fa:16:3e:96:46:60", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbd71fc-00", "ovs_interfaceid": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "address": "fa:16:3e:f5:db:69", "network": {"id": "3a407f9b-a9f0-4839-8dd2-f10ecd904b43", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1652315845", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda35a09f-ea", "ovs_interfaceid": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b569b424-48e7-45f4-8f0d-53192e53abf0", "address": "fa:16:3e:53:7c:52", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.192", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb569b424-48", "ovs_interfaceid": "b569b424-48e7-45f4-8f0d-53192e53abf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.480552] env[61995]: DEBUG nova.scheduler.client.report [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.495103] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794283, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.507221] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794282, 'name': ReconfigVM_Task, 'duration_secs': 0.771323} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.507492] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.508124] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91baf397-b6a8-494f-bb12-c6c9c65b9a33 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.517431] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 719.517431] env[61995]: value = "task-794284" [ 719.517431] env[61995]: _type = "Task" [ 719.517431] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.524851] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794284, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.736268] env[61995]: DEBUG nova.network.neutron [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.769357] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e05c68-f07f-bdca-644f-36ef7e63dcfb, 'name': SearchDatastore_Task, 'duration_secs': 0.00978} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.772270] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b192df2e-0b31-4536-87bb-524138051731 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.777627] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 719.777627] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525ff47e-c67e-d2ef-18f7-6e9121137958" [ 719.777627] env[61995]: _type = "Task" [ 719.777627] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.785263] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525ff47e-c67e-d2ef-18f7-6e9121137958, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.881744] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5229c3ac-09b8-1aa7-e488-ccd776620b10, 'name': SearchDatastore_Task, 'duration_secs': 0.00861} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.882551] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6553733-0e40-40eb-a35f-770043f9e28a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.887588] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 719.887588] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522915fb-3ffb-69cb-1faf-29eaf61bdfa8" [ 719.887588] env[61995]: _type = "Task" [ 719.887588] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.892893] env[61995]: DEBUG oslo_concurrency.lockutils [req-a1ce23dd-cf68-4c68-8576-83ca64b2e116 req-394d2a3a-f186-466f-8a5d-22e3ba28923c service nova] Releasing lock "refresh_cache-1c4eb398-c4e9-4588-ba48-2805e3fee750" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.898208] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522915fb-3ffb-69cb-1faf-29eaf61bdfa8, 'name': SearchDatastore_Task, 'duration_secs': 0.008443} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.898434] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.898676] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] b1bd98d9-bd0f-4abd-a188-e5267ada4852/b1bd98d9-bd0f-4abd-a188-e5267ada4852.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 719.898920] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3517cbab-e3d1-44fc-9a7a-04b17039865d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.906067] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 719.906067] env[61995]: value = "task-794285" [ 719.906067] env[61995]: _type = "Task" [ 719.906067] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.913519] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794285, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.942707] env[61995]: DEBUG nova.network.neutron [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Updating instance_info_cache with network_info: [{"id": "f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9", "address": "fa:16:3e:33:2e:ec", "network": {"id": "05b3b850-c61b-42fe-897a-6a21fee9311b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1871881368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64e0bdabc3cf4eb7b606adda78f9a0b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf97a3a6e-f6", "ovs_interfaceid": "f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.986078] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.987709] env[61995]: DEBUG nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.993025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.047s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.993261] env[61995]: DEBUG nova.objects.instance [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lazy-loading 'resources' on Instance uuid d5a54c46-1a45-4bd5-8694-0defc757c455 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 720.005575] env[61995]: DEBUG oslo_vmware.api [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794283, 'name': PowerOnVM_Task, 'duration_secs': 0.784679} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.006432] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 720.006675] env[61995]: INFO nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Took 10.17 seconds to spawn the instance on the hypervisor. [ 720.006856] env[61995]: DEBUG nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 720.007651] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6374b8-4d9d-4b4e-8efe-e948af69ae3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.025611] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794284, 'name': Rename_Task, 'duration_secs': 0.25572} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.026337] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.026337] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec5e5616-98c4-4d27-80db-36bfa0d51a47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.035329] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 720.035329] env[61995]: value = "task-794286" [ 720.035329] env[61995]: _type = "Task" [ 720.035329] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.048599] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794286, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.141818] env[61995]: DEBUG nova.network.neutron [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updated VIF entry in instance network info cache for port 802b752a-238e-4b55-8bd3-f73eddc4b0f4. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 720.142259] env[61995]: DEBUG nova.network.neutron [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.291449] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525ff47e-c67e-d2ef-18f7-6e9121137958, 'name': SearchDatastore_Task, 'duration_secs': 0.00921} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.291789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.292135] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 1c4eb398-c4e9-4588-ba48-2805e3fee750/1c4eb398-c4e9-4588-ba48-2805e3fee750.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 720.292479] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21b2d86e-4d63-4b76-9c0d-7acb7f410370 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.301678] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 720.301678] env[61995]: value = "task-794287" [ 720.301678] env[61995]: _type = "Task" [ 720.301678] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.314108] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794287, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.416653] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794285, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490697} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.416985] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] b1bd98d9-bd0f-4abd-a188-e5267ada4852/b1bd98d9-bd0f-4abd-a188-e5267ada4852.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 720.417224] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 720.417472] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13481477-6a47-41dd-bb1d-949e0fd7287d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.424530] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 720.424530] env[61995]: value = "task-794288" [ 720.424530] env[61995]: _type = "Task" [ 720.424530] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.435060] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794288, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.445845] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Releasing lock "refresh_cache-19fd0358-32e4-434f-b0b3-c8cc6c07ba51" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.446227] env[61995]: DEBUG nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Instance network_info: |[{"id": "f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9", "address": "fa:16:3e:33:2e:ec", "network": {"id": "05b3b850-c61b-42fe-897a-6a21fee9311b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1871881368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64e0bdabc3cf4eb7b606adda78f9a0b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf97a3a6e-f6", "ovs_interfaceid": "f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 720.446545] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] Acquired lock "refresh_cache-19fd0358-32e4-434f-b0b3-c8cc6c07ba51" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.446726] env[61995]: DEBUG nova.network.neutron [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Refreshing network info cache for port f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.448060] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:2e:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.455574] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Creating folder: Project (64e0bdabc3cf4eb7b606adda78f9a0b4). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 720.456141] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-205d9ff6-30fa-4da4-be6e-ccd1c2ce179d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.468341] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Created folder: Project (64e0bdabc3cf4eb7b606adda78f9a0b4) in parent group-v185203. [ 720.468531] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Creating folder: Instances. Parent ref: group-v185240. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 720.468768] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86f98c86-e463-44e0-a474-e2b25381806c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.478354] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Created folder: Instances in parent group-v185240. [ 720.478591] env[61995]: DEBUG oslo.service.loopingcall [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.478781] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 720.479027] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ec441bb1-e1fe-424c-a372-1ea801dbbadb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.499137] env[61995]: DEBUG nova.compute.utils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.504166] env[61995]: DEBUG nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.504376] env[61995]: DEBUG nova.network.neutron [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 720.506300] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.506300] env[61995]: value = "task-794291" [ 720.506300] env[61995]: _type = "Task" [ 720.506300] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.518113] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794291, 'name': CreateVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.530470] env[61995]: INFO nova.compute.manager [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Took 30.68 seconds to build instance. [ 720.546693] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794286, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.554090] env[61995]: DEBUG nova.policy [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2ec99435d404ea1a22caadbdfd2ed5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da64cb2aceb445a3b13d15376de7f852', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 720.646685] env[61995]: DEBUG oslo_concurrency.lockutils [req-f7e471ff-6ee1-4f0a-98ae-816d4cf36535 req-1e460479-7204-45f8-a3d3-f69175d13877 service nova] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.821098] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794287, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.940737] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794288, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065749} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.941248] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 720.941812] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f758d5c-a554-4959-9dad-e0c58e48c0f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.964657] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Reconfiguring VM instance instance-00000020 to attach disk [datastore1] b1bd98d9-bd0f-4abd-a188-e5267ada4852/b1bd98d9-bd0f-4abd-a188-e5267ada4852.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 720.970297] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b9e5f9e-4e9e-41a5-b96e-3c33b0dd95b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.991082] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 720.991082] env[61995]: value = "task-794292" [ 720.991082] env[61995]: _type = "Task" [ 720.991082] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.999273] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794292, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.010675] env[61995]: DEBUG nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.025091] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794291, 'name': CreateVM_Task, 'duration_secs': 0.407732} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.025268] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 721.025925] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.026105] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.026429] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 721.026682] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bd04784-bb8d-4971-9ebd-e1b7850f96ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.031162] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 721.031162] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5273c8b0-750a-ff71-5730-9f8b01a93ca9" [ 721.031162] env[61995]: _type = "Task" [ 721.031162] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.034434] env[61995]: DEBUG oslo_concurrency.lockutils [None req-640625b5-e766-43ac-950e-f4e9c47e7f18 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.242s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.044897] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5273c8b0-750a-ff71-5730-9f8b01a93ca9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.045998] env[61995]: DEBUG nova.network.neutron [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Successfully created port: ef74f12f-3ba8-4392-b066-245f7c8a6425 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.053804] env[61995]: DEBUG oslo_vmware.api [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794286, 'name': PowerOnVM_Task, 'duration_secs': 0.551114} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.056210] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.056441] env[61995]: INFO nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Took 8.69 seconds to spawn the instance on the hypervisor. [ 721.056587] env[61995]: DEBUG nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.058183] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2db82f1-dc68-4cc8-aaa3-a018c95ff97a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.106800] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849d237b-d025-4c60-9e7e-951de6695524 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.115970] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdf6d8b-5b45-4d7a-b00d-00d81b912a44 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.151813] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58dcd5aa-6a51-4915-84ad-f43122c428fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.159802] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b24fc8-bade-4b74-91ef-b1892feb77ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.173067] env[61995]: DEBUG nova.compute.provider_tree [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.315706] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794287, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567076} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.315706] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 1c4eb398-c4e9-4588-ba48-2805e3fee750/1c4eb398-c4e9-4588-ba48-2805e3fee750.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 721.315706] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.315706] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0db897a2-5d80-4ba0-a7c0-9d324903ee30 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.321854] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 721.321854] env[61995]: value = "task-794293" [ 721.321854] env[61995]: _type = "Task" [ 721.321854] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.332658] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.432362] env[61995]: DEBUG nova.network.neutron [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Updated VIF entry in instance network info cache for port f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 721.432717] env[61995]: DEBUG nova.network.neutron [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Updating instance_info_cache with network_info: [{"id": "f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9", "address": "fa:16:3e:33:2e:ec", "network": {"id": "05b3b850-c61b-42fe-897a-6a21fee9311b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1871881368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64e0bdabc3cf4eb7b606adda78f9a0b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf97a3a6e-f6", "ovs_interfaceid": "f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.503899] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794292, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.543506] env[61995]: DEBUG nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.545758] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5273c8b0-750a-ff71-5730-9f8b01a93ca9, 'name': SearchDatastore_Task, 'duration_secs': 0.009121} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.549021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.549021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 721.549021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.549021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.549021] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 721.549021] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dc9f816-3386-4e59-a2a7-e27d818355ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.555317] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 721.555483] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 721.556202] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4221c25-73b6-4193-af24-57cc11003261 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.562318] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 721.562318] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5237123c-d341-417c-173b-15f042293781" [ 721.562318] env[61995]: _type = "Task" [ 721.562318] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.574930] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5237123c-d341-417c-173b-15f042293781, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.580232] env[61995]: INFO nova.compute.manager [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Took 28.83 seconds to build instance. [ 721.594819] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.597670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.597670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.597670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.597670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.602175] env[61995]: INFO nova.compute.manager [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Terminating instance [ 721.602874] env[61995]: DEBUG nova.compute.manager [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 721.603181] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 721.604375] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4978d9-5ab7-415e-822d-54d7ac34fe69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.612223] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 721.612634] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50a764c1-139c-4222-89a2-4bf495050c3c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.618560] env[61995]: DEBUG oslo_vmware.api [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 721.618560] env[61995]: value = "task-794294" [ 721.618560] env[61995]: _type = "Task" [ 721.618560] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.634294] env[61995]: DEBUG oslo_vmware.api [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.702894] env[61995]: ERROR nova.scheduler.client.report [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] [req-ced1b907-f998-4fd4-bb22-124a1a5bbee1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ced1b907-f998-4fd4-bb22-124a1a5bbee1"}]} [ 721.725575] env[61995]: DEBUG nova.scheduler.client.report [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 721.751403] env[61995]: DEBUG nova.scheduler.client.report [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 721.751403] env[61995]: DEBUG nova.compute.provider_tree [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.766394] env[61995]: DEBUG nova.scheduler.client.report [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 721.788418] env[61995]: DEBUG nova.scheduler.client.report [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 721.833749] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.274464} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.836838] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.837852] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183f55c7-0401-4a58-9af2-fb2ace831400 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.865297] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] 1c4eb398-c4e9-4588-ba48-2805e3fee750/1c4eb398-c4e9-4588-ba48-2805e3fee750.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.868278] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be9704d0-87d9-4092-a104-5f318974b7aa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.892801] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 721.892801] env[61995]: value = "task-794295" [ 721.892801] env[61995]: _type = "Task" [ 721.892801] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.899981] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.936159] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a4c16a0-d912-4d51-af82-901a3a82a4ce req-7085a6e1-3862-48ab-bbdf-613c40a106fc service nova] Releasing lock "refresh_cache-19fd0358-32e4-434f-b0b3-c8cc6c07ba51" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.003797] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794292, 'name': ReconfigVM_Task, 'duration_secs': 0.872304} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.004209] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Reconfigured VM instance instance-00000020 to attach disk [datastore1] b1bd98d9-bd0f-4abd-a188-e5267ada4852/b1bd98d9-bd0f-4abd-a188-e5267ada4852.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.005220] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f70b6445-3691-41a4-a768-c05d05603ab4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.014173] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 722.014173] env[61995]: value = "task-794296" [ 722.014173] env[61995]: _type = "Task" [ 722.014173] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.030360] env[61995]: DEBUG nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.032162] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794296, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.064068] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.064364] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.064557] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.064766] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.064941] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.065134] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.065402] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.065602] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.065805] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.066051] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.066276] env[61995]: DEBUG nova.virt.hardware [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.067227] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2492b7f1-e6d6-4b17-9ad9-a709f0728866 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.076904] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.085643] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88583188-a68b-4780-b78a-7fa9bc804c41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.089943] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59fa135e-7751-4af1-99a4-dacb34fc4058 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.466s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.090516] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5237123c-d341-417c-173b-15f042293781, 'name': SearchDatastore_Task, 'duration_secs': 0.008552} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.095661] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5956f0cf-9698-41fe-a75c-69d0026d7138 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.115131] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 722.115131] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520be1de-b628-1fc8-7629-8c1ef034483c" [ 722.115131] env[61995]: _type = "Task" [ 722.115131] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.128533] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520be1de-b628-1fc8-7629-8c1ef034483c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.133578] env[61995]: DEBUG oslo_vmware.api [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794294, 'name': PowerOffVM_Task, 'duration_secs': 0.253129} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.133831] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 722.133998] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 722.134273] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89d93873-ffe3-4d2a-9b9f-c52249729a94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.228364] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 722.228590] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 722.228749] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Deleting the datastore file [datastore1] 24db5fad-52fd-4689-a8aa-2ae4b0a06710 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 722.229942] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a757f58a-df38-4421-89b0-0ef71e2b59dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.235376] env[61995]: DEBUG oslo_vmware.api [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for the task: (returnval){ [ 722.235376] env[61995]: value = "task-794298" [ 722.235376] env[61995]: _type = "Task" [ 722.235376] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.246956] env[61995]: DEBUG oslo_vmware.api [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.277199] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d1f4cd-d296-4a98-aa4d-eafbe939a997 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.285891] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4712d378-537b-47ef-82dc-cf650c4bbada {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.316727] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1b2cf8-fbeb-4d45-936c-5c2e6dee5c1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.324532] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2c7e54-398f-4540-8fcb-94df0d835fd3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.339745] env[61995]: DEBUG nova.compute.provider_tree [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.400408] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.525145] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794296, 'name': Rename_Task, 'duration_secs': 0.270879} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.525494] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 722.525741] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a5ed811-fc1c-462c-a5ed-08d325d11836 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.531969] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 722.531969] env[61995]: value = "task-794299" [ 722.531969] env[61995]: _type = "Task" [ 722.531969] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.540050] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.599728] env[61995]: DEBUG nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.625550] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520be1de-b628-1fc8-7629-8c1ef034483c, 'name': SearchDatastore_Task, 'duration_secs': 0.010131} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.625815] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.626084] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 19fd0358-32e4-434f-b0b3-c8cc6c07ba51/19fd0358-32e4-434f-b0b3-c8cc6c07ba51.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 722.626342] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6d44c2d-64c6-4430-93e1-ed49dfc493e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.633098] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 722.633098] env[61995]: value = "task-794300" [ 722.633098] env[61995]: _type = "Task" [ 722.633098] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.641753] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794300, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.694851] env[61995]: INFO nova.compute.manager [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Rebuilding instance [ 722.744990] env[61995]: DEBUG nova.compute.manager [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 722.745798] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61211cd2-d5dd-473c-9555-310936124d17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.752233] env[61995]: DEBUG oslo_vmware.api [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Task: {'id': task-794298, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139883} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.752830] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 722.753105] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 722.753309] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 722.753359] env[61995]: INFO nova.compute.manager [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Took 1.15 seconds to destroy the instance on the hypervisor. [ 722.753739] env[61995]: DEBUG oslo.service.loopingcall [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.753951] env[61995]: DEBUG nova.compute.manager [-] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.754071] env[61995]: DEBUG nova.network.neutron [-] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 722.901472] env[61995]: DEBUG nova.scheduler.client.report [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 55 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 722.901776] env[61995]: DEBUG nova.compute.provider_tree [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 55 to 56 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 722.901952] env[61995]: DEBUG nova.compute.provider_tree [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.907560] env[61995]: DEBUG nova.network.neutron [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Successfully updated port: ef74f12f-3ba8-4392-b066-245f7c8a6425 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 722.915547] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794295, 'name': ReconfigVM_Task, 'duration_secs': 0.56332} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.917152] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Reconfigured VM instance instance-0000001d to attach disk [datastore2] 1c4eb398-c4e9-4588-ba48-2805e3fee750/1c4eb398-c4e9-4588-ba48-2805e3fee750.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.918655] env[61995]: DEBUG nova.compute.manager [req-7d2d7e7d-d4db-4e68-ac6a-847d2fc8cf1e req-69201677-6db7-4624-9c7d-b2eaf3fe962d service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Received event network-vif-plugged-ef74f12f-3ba8-4392-b066-245f7c8a6425 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 722.919066] env[61995]: DEBUG oslo_concurrency.lockutils [req-7d2d7e7d-d4db-4e68-ac6a-847d2fc8cf1e req-69201677-6db7-4624-9c7d-b2eaf3fe962d service nova] Acquiring lock "c350194c-ed3e-4996-871a-70b4cb99c7db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.919569] env[61995]: DEBUG oslo_concurrency.lockutils [req-7d2d7e7d-d4db-4e68-ac6a-847d2fc8cf1e req-69201677-6db7-4624-9c7d-b2eaf3fe962d service nova] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.921706] env[61995]: DEBUG oslo_concurrency.lockutils [req-7d2d7e7d-d4db-4e68-ac6a-847d2fc8cf1e req-69201677-6db7-4624-9c7d-b2eaf3fe962d service nova] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.921706] env[61995]: DEBUG nova.compute.manager [req-7d2d7e7d-d4db-4e68-ac6a-847d2fc8cf1e req-69201677-6db7-4624-9c7d-b2eaf3fe962d service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] No waiting events found dispatching network-vif-plugged-ef74f12f-3ba8-4392-b066-245f7c8a6425 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 722.921706] env[61995]: WARNING nova.compute.manager [req-7d2d7e7d-d4db-4e68-ac6a-847d2fc8cf1e req-69201677-6db7-4624-9c7d-b2eaf3fe962d service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Received unexpected event network-vif-plugged-ef74f12f-3ba8-4392-b066-245f7c8a6425 for instance with vm_state building and task_state spawning. [ 722.921706] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42365c97-2746-4a77-82b6-5c03ddd82d6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.930295] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 722.930295] env[61995]: value = "task-794301" [ 722.930295] env[61995]: _type = "Task" [ 722.930295] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.942814] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794301, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.043446] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794299, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.126530] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.145235] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794300, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455476} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.145614] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 19fd0358-32e4-434f-b0b3-c8cc6c07ba51/19fd0358-32e4-434f-b0b3-c8cc6c07ba51.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 723.145739] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.145968] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c7ed843-b4f3-4118-ac97-93a29e3e3046 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.152412] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 723.152412] env[61995]: value = "task-794302" [ 723.152412] env[61995]: _type = "Task" [ 723.152412] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.161762] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794302, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.204388] env[61995]: DEBUG nova.compute.manager [req-094ac7b0-7cd6-4e71-92ef-9b7b1e60529d req-2d3f3936-e9e8-4b06-9fe5-a39f39adf3d7 service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Received event network-vif-deleted-1999392b-e938-43fc-9361-7179466a8aea {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 723.204586] env[61995]: INFO nova.compute.manager [req-094ac7b0-7cd6-4e71-92ef-9b7b1e60529d req-2d3f3936-e9e8-4b06-9fe5-a39f39adf3d7 service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Neutron deleted interface 1999392b-e938-43fc-9361-7179466a8aea; detaching it from the instance and deleting it from the info cache [ 723.204798] env[61995]: DEBUG nova.network.neutron [req-094ac7b0-7cd6-4e71-92ef-9b7b1e60529d req-2d3f3936-e9e8-4b06-9fe5-a39f39adf3d7 service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.262065] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 723.262490] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74cc7528-4253-43fd-83f7-234510905fd6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.270255] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 723.270255] env[61995]: value = "task-794303" [ 723.270255] env[61995]: _type = "Task" [ 723.270255] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.279079] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.410722] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "refresh_cache-c350194c-ed3e-4996-871a-70b4cb99c7db" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.410722] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquired lock "refresh_cache-c350194c-ed3e-4996-871a-70b4cb99c7db" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.410864] env[61995]: DEBUG nova.network.neutron [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.412462] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.420s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.415029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.048s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.440774] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794301, 'name': Rename_Task, 'duration_secs': 0.192628} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.441795] env[61995]: INFO nova.scheduler.client.report [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Deleted allocations for instance d5a54c46-1a45-4bd5-8694-0defc757c455 [ 723.443030] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 723.445341] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff99ed57-100f-4b48-a5e6-2533bc7a13f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.452760] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 723.452760] env[61995]: value = "task-794304" [ 723.452760] env[61995]: _type = "Task" [ 723.452760] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.460556] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794304, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.543442] env[61995]: DEBUG oslo_vmware.api [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794299, 'name': PowerOnVM_Task, 'duration_secs': 0.53024} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.543833] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 723.544099] env[61995]: INFO nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Took 8.68 seconds to spawn the instance on the hypervisor. [ 723.544320] env[61995]: DEBUG nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 723.545799] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa839cdb-461f-48bc-8dfe-cdc6131e1800 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.641842] env[61995]: DEBUG nova.network.neutron [-] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.661387] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794302, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.425291} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.661641] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.662409] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d5053a-446e-47a1-882d-5bb4891b4918 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.685944] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 19fd0358-32e4-434f-b0b3-c8cc6c07ba51/19fd0358-32e4-434f-b0b3-c8cc6c07ba51.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.686565] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-876ce984-c31e-4ff0-9f24-ac5213410f17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.707681] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 723.707681] env[61995]: value = "task-794305" [ 723.707681] env[61995]: _type = "Task" [ 723.707681] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.710883] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c68e0ca-8a58-4469-8be3-c2c3f050377d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.719339] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794305, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.727021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9582852a-8876-44d6-9e47-ed10a30ad0b6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.754189] env[61995]: DEBUG nova.compute.manager [req-094ac7b0-7cd6-4e71-92ef-9b7b1e60529d req-2d3f3936-e9e8-4b06-9fe5-a39f39adf3d7 service nova] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Detach interface failed, port_id=1999392b-e938-43fc-9361-7179466a8aea, reason: Instance 24db5fad-52fd-4689-a8aa-2ae4b0a06710 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 723.779714] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794303, 'name': PowerOffVM_Task, 'duration_secs': 0.431132} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.780009] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 723.780259] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 723.781066] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a22888c-aae4-4f6c-9185-dd569b2186bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.789022] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 723.789022] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a68ba57-35f5-418e-8918-2e2e8f1861a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.864323] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 723.864597] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 723.864795] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleting the datastore file [datastore1] 45fdae21-bec2-438a-8f2a-c23b0e428057 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 723.865097] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-025fb146-9fe1-4644-aba4-5b906aca4eb7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.871830] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 723.871830] env[61995]: value = "task-794307" [ 723.871830] env[61995]: _type = "Task" [ 723.871830] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.880239] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.952994] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b245ced6-33c6-43a7-824d-3e266725719d tempest-ListImageFiltersTestJSON-708913367 tempest-ListImageFiltersTestJSON-708913367-project-member] Lock "d5a54c46-1a45-4bd5-8694-0defc757c455" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.039s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.965406] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794304, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.967176] env[61995]: DEBUG nova.network.neutron [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.065318] env[61995]: INFO nova.compute.manager [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Took 29.55 seconds to build instance. [ 724.142885] env[61995]: INFO nova.compute.manager [-] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Took 1.39 seconds to deallocate network for instance. [ 724.148367] env[61995]: DEBUG nova.network.neutron [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Updating instance_info_cache with network_info: [{"id": "ef74f12f-3ba8-4392-b066-245f7c8a6425", "address": "fa:16:3e:86:3d:23", "network": {"id": "45251e67-694d-49fc-800e-4d363a168df6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-884649296-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da64cb2aceb445a3b13d15376de7f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e23c1d18-c841-49ea-95f3-df5ceac28afd", "external-id": "nsx-vlan-transportzone-774", "segmentation_id": 774, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef74f12f-3b", "ovs_interfaceid": "ef74f12f-3ba8-4392-b066-245f7c8a6425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.218723] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794305, 'name': ReconfigVM_Task, 'duration_secs': 0.290107} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.219075] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 19fd0358-32e4-434f-b0b3-c8cc6c07ba51/19fd0358-32e4-434f-b0b3-c8cc6c07ba51.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.219768] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d39785ab-8c40-4cc7-a1b9-1b37dddca73b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.226020] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 724.226020] env[61995]: value = "task-794308" [ 724.226020] env[61995]: _type = "Task" [ 724.226020] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.233804] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794308, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.384031] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12873} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.384306] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 724.384493] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 724.384671] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 724.447641] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e78f8718-e0be-4b2a-ab70-9565ce01ab9d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 724.447831] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance f47e5bbb-f4a8-4686-8195-8348d2ffb08d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 724.448038] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 08489e01-2ce3-45a1-8f87-2a4c89230aa2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 724.448225] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e53d10a3-1d16-498d-b4d8-abe7205410ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.448297] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d9c53806-68c0-4872-a262-36bc05573674 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.448406] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 1c4eb398-c4e9-4588-ba48-2805e3fee750 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.448515] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 24db5fad-52fd-4689-a8aa-2ae4b0a06710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.448621] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 45fdae21-bec2-438a-8f2a-c23b0e428057 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.448723] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b1bd98d9-bd0f-4abd-a188-e5267ada4852 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.448828] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 19fd0358-32e4-434f-b0b3-c8cc6c07ba51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.450049] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance c350194c-ed3e-4996-871a-70b4cb99c7db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.465441] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794304, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.569033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-40fa8269-e1d0-4bc8-a6cd-4bdc759e36fe tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.580s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.651044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.651509] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Releasing lock "refresh_cache-c350194c-ed3e-4996-871a-70b4cb99c7db" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.651839] env[61995]: DEBUG nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Instance network_info: |[{"id": "ef74f12f-3ba8-4392-b066-245f7c8a6425", "address": "fa:16:3e:86:3d:23", "network": {"id": "45251e67-694d-49fc-800e-4d363a168df6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-884649296-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da64cb2aceb445a3b13d15376de7f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e23c1d18-c841-49ea-95f3-df5ceac28afd", "external-id": "nsx-vlan-transportzone-774", "segmentation_id": 774, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef74f12f-3b", "ovs_interfaceid": "ef74f12f-3ba8-4392-b066-245f7c8a6425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 724.652960] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:3d:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e23c1d18-c841-49ea-95f3-df5ceac28afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef74f12f-3ba8-4392-b066-245f7c8a6425', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.660270] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Creating folder: Project (da64cb2aceb445a3b13d15376de7f852). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.660569] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7740f3e2-b4b4-4c3d-8794-9bc9cc8b46b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.670793] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Created folder: Project (da64cb2aceb445a3b13d15376de7f852) in parent group-v185203. [ 724.670978] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Creating folder: Instances. Parent ref: group-v185243. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.671224] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd4509fb-01b0-4be7-bb1d-d09830a77ae1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.679459] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Created folder: Instances in parent group-v185243. [ 724.679628] env[61995]: DEBUG oslo.service.loopingcall [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.679822] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 724.680057] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f47d578-871e-4a61-a95a-32b7e7f5956e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.698957] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.698957] env[61995]: value = "task-794311" [ 724.698957] env[61995]: _type = "Task" [ 724.698957] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.708446] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794311, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.735076] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794308, 'name': Rename_Task, 'duration_secs': 0.175333} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.735417] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 724.735590] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b46a2393-d0c0-46b4-948a-cb24c2e06d53 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.743416] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 724.743416] env[61995]: value = "task-794312" [ 724.743416] env[61995]: _type = "Task" [ 724.743416] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.754152] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794312, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.952927] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance a04fbe7e-e53e-43ef-9a6f-ba3895ce7597 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 724.965327] env[61995]: DEBUG oslo_vmware.api [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794304, 'name': PowerOnVM_Task, 'duration_secs': 1.048125} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.965587] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 724.965784] env[61995]: INFO nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Took 17.67 seconds to spawn the instance on the hypervisor. [ 724.965987] env[61995]: DEBUG nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 724.966790] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843cfe3a-60cf-4f3d-be5e-5599646f6238 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.016393] env[61995]: DEBUG nova.compute.manager [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Received event network-changed-ef74f12f-3ba8-4392-b066-245f7c8a6425 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.016393] env[61995]: DEBUG nova.compute.manager [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Refreshing instance network info cache due to event network-changed-ef74f12f-3ba8-4392-b066-245f7c8a6425. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 725.016393] env[61995]: DEBUG oslo_concurrency.lockutils [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] Acquiring lock "refresh_cache-c350194c-ed3e-4996-871a-70b4cb99c7db" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.016393] env[61995]: DEBUG oslo_concurrency.lockutils [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] Acquired lock "refresh_cache-c350194c-ed3e-4996-871a-70b4cb99c7db" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.016393] env[61995]: DEBUG nova.network.neutron [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Refreshing network info cache for port ef74f12f-3ba8-4392-b066-245f7c8a6425 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.071376] env[61995]: DEBUG nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.210082] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794311, 'name': CreateVM_Task, 'duration_secs': 0.387656} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.210454] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 725.211155] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.211353] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.211679] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 725.211933] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-997d9945-87e0-4d3a-b258-733a702f9abc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.218625] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 725.218625] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c9ba16-4ee1-d677-c7b9-b046456636c6" [ 725.218625] env[61995]: _type = "Task" [ 725.218625] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.229247] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c9ba16-4ee1-d677-c7b9-b046456636c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.253776] env[61995]: DEBUG oslo_vmware.api [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794312, 'name': PowerOnVM_Task, 'duration_secs': 0.4921} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.255048] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.255359] env[61995]: INFO nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Took 7.88 seconds to spawn the instance on the hypervisor. [ 725.255550] env[61995]: DEBUG nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.256393] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e12cf20-21ec-4e81-9e97-fc1b3e58dcb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.433292] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.434550] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.434550] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.436409] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.436409] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.436409] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.436409] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.436409] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.436409] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.436409] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.436774] env[61995]: DEBUG nova.virt.hardware [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.438033] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07db4ef4-7e92-4b18-9b74-e2196350c20b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.446628] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee22820e-0465-429e-8219-6b2bb65ef857 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.465379] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance c4f744f9-36f3-4913-9ced-f0db93d3b4df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 725.467015] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:7f:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '515eb152-ff0c-4c68-896f-d46c12f8c848', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.477269] env[61995]: DEBUG oslo.service.loopingcall [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.479412] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 725.485808] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-990f5b5d-134d-4253-8d79-b467442fa289 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.504601] env[61995]: INFO nova.compute.manager [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Took 42.83 seconds to build instance. [ 725.511290] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.511290] env[61995]: value = "task-794313" [ 725.511290] env[61995]: _type = "Task" [ 725.511290] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.522829] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794313, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.598358] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.729061] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c9ba16-4ee1-d677-c7b9-b046456636c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010114} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.729600] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.730084] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.730422] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.730670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.730957] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.731671] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9cdf129e-e7aa-4b0e-99f1-45c08bf51786 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.745059] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.745059] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 725.745059] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ae685c4-52c5-49a5-9556-db2ec0eb11b6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.752028] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 725.752028] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52370b6c-3686-3b16-c605-10e1a52735fe" [ 725.752028] env[61995]: _type = "Task" [ 725.752028] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.761722] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52370b6c-3686-3b16-c605-10e1a52735fe, 'name': SearchDatastore_Task, 'duration_secs': 0.008218} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.762782] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1552d3-9db2-4447-bcfa-185c06aece74 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.777894] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 725.777894] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5279f265-aa33-ad2b-132f-2d05c7ded70a" [ 725.777894] env[61995]: _type = "Task" [ 725.777894] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.779671] env[61995]: INFO nova.compute.manager [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Took 30.58 seconds to build instance. [ 725.790460] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5279f265-aa33-ad2b-132f-2d05c7ded70a, 'name': SearchDatastore_Task, 'duration_secs': 0.009038} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.790719] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.790965] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c350194c-ed3e-4996-871a-70b4cb99c7db/c350194c-ed3e-4996-871a-70b4cb99c7db.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 725.791257] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67e28c85-b71c-4a69-8b54-3ecf7c8c41d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.799553] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 725.799553] env[61995]: value = "task-794314" [ 725.799553] env[61995]: _type = "Task" [ 725.799553] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.803630] env[61995]: DEBUG nova.network.neutron [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Updated VIF entry in instance network info cache for port ef74f12f-3ba8-4392-b066-245f7c8a6425. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 725.803970] env[61995]: DEBUG nova.network.neutron [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Updating instance_info_cache with network_info: [{"id": "ef74f12f-3ba8-4392-b066-245f7c8a6425", "address": "fa:16:3e:86:3d:23", "network": {"id": "45251e67-694d-49fc-800e-4d363a168df6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-884649296-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da64cb2aceb445a3b13d15376de7f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e23c1d18-c841-49ea-95f3-df5ceac28afd", "external-id": "nsx-vlan-transportzone-774", "segmentation_id": 774, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef74f12f-3b", "ovs_interfaceid": "ef74f12f-3ba8-4392-b066-245f7c8a6425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.810815] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794314, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.979197] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance fcdc7c33-8c46-4f4d-bb39-e6f943256cb0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 726.007371] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1454e52-ffb1-4eee-9905-e0db8d35e41c tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.386s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.030071] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794313, 'name': CreateVM_Task, 'duration_secs': 0.372786} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.030951] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 726.031707] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.031882] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.032241] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 726.033281] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee748827-91d5-4945-936d-670842c6f0d8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.039587] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 726.039587] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f665c1-b011-d359-ed91-12f29b908351" [ 726.039587] env[61995]: _type = "Task" [ 726.039587] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.051437] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f665c1-b011-d359-ed91-12f29b908351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.285526] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57928cf5-a052-44b2-abcc-0c838c9695a6 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.279s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.306932] env[61995]: DEBUG oslo_concurrency.lockutils [req-3da0b3f6-cf9d-4a90-a454-db43534c0e2c req-5b5b42b2-6f01-48fd-a9e3-73ea11e5042c service nova] Releasing lock "refresh_cache-c350194c-ed3e-4996-871a-70b4cb99c7db" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.310912] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794314, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496526} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.311164] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c350194c-ed3e-4996-871a-70b4cb99c7db/c350194c-ed3e-4996-871a-70b4cb99c7db.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 726.311373] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.311617] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f60a0e2c-b3ad-4167-92b8-60eca6f8a3ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.317631] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 726.317631] env[61995]: value = "task-794315" [ 726.317631] env[61995]: _type = "Task" [ 726.317631] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.326072] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794315, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.420347] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "1c4eb398-c4e9-4588-ba48-2805e3fee750" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.420687] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.420902] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.421102] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.421278] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.423904] env[61995]: INFO nova.compute.manager [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Terminating instance [ 726.427308] env[61995]: DEBUG nova.compute.manager [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.427508] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 726.428374] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68023fa9-e4bb-44c5-833c-b44d739e0e47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.436610] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 726.436843] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d597e6c7-d890-48b3-bb76-4fae1256ff32 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.443490] env[61995]: DEBUG oslo_vmware.api [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 726.443490] env[61995]: value = "task-794316" [ 726.443490] env[61995]: _type = "Task" [ 726.443490] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.451401] env[61995]: DEBUG oslo_vmware.api [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794316, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.485038] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 55555deb-9883-483b-a901-42b1a66c60d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 726.510476] env[61995]: DEBUG nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.551052] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f665c1-b011-d359-ed91-12f29b908351, 'name': SearchDatastore_Task, 'duration_secs': 0.02372} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.552074] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.552380] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.552625] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.552773] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.552951] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.553473] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5efb0ed7-b3e3-4916-9359-5f2368186014 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.563226] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.563226] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 726.563714] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c01728b-f849-4815-815c-fe747155a959 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.569666] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 726.569666] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52780e46-0832-e640-f9c0-fa88442602e6" [ 726.569666] env[61995]: _type = "Task" [ 726.569666] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.577484] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52780e46-0832-e640-f9c0-fa88442602e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.788055] env[61995]: DEBUG nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.831589] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794315, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063394} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.831856] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 726.832646] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce4131b-5706-4ad2-affb-f1a0c84a915e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.855698] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] c350194c-ed3e-4996-871a-70b4cb99c7db/c350194c-ed3e-4996-871a-70b4cb99c7db.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 726.856252] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b1bd1a5-6a26-40e0-93c4-7ef3e05dab67 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.878863] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 726.878863] env[61995]: value = "task-794317" [ 726.878863] env[61995]: _type = "Task" [ 726.878863] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.886973] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794317, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.955263] env[61995]: DEBUG oslo_vmware.api [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794316, 'name': PowerOffVM_Task, 'duration_secs': 0.410124} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.955491] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 726.955663] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 726.955935] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86af5a7d-2702-4712-b431-9a9d2009886a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.988855] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ecb72668-caa4-4117-9f8d-e0c6e4831f3d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 727.036532] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.083512] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52780e46-0832-e640-f9c0-fa88442602e6, 'name': SearchDatastore_Task, 'duration_secs': 0.008483} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.084744] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a411d7fd-8327-4c8f-975a-4b4f0933cd39 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.092035] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 727.092035] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a0b88b-3f39-b15f-b78d-8c2565767d11" [ 727.092035] env[61995]: _type = "Task" [ 727.092035] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.102697] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a0b88b-3f39-b15f-b78d-8c2565767d11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.161021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 727.161021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 727.161021] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Deleting the datastore file [datastore2] 1c4eb398-c4e9-4588-ba48-2805e3fee750 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 727.161021] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26f5ecaa-8659-4160-8e3a-c88bc527777e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.170135] env[61995]: DEBUG oslo_vmware.api [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 727.170135] env[61995]: value = "task-794319" [ 727.170135] env[61995]: _type = "Task" [ 727.170135] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.172783] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dd7576c3-7cba-4c16-ab91-92b738615eec tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "interface-19fd0358-32e4-434f-b0b3-c8cc6c07ba51-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.173436] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dd7576c3-7cba-4c16-ab91-92b738615eec tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "interface-19fd0358-32e4-434f-b0b3-c8cc6c07ba51-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.174030] env[61995]: DEBUG nova.objects.instance [None req-dd7576c3-7cba-4c16-ab91-92b738615eec tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lazy-loading 'flavor' on Instance uuid 19fd0358-32e4-434f-b0b3-c8cc6c07ba51 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.182584] env[61995]: DEBUG oslo_vmware.api [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.207379] env[61995]: DEBUG nova.compute.manager [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received event network-changed-0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.207379] env[61995]: DEBUG nova.compute.manager [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Refreshing instance network info cache due to event network-changed-0f5a4362-281d-470a-95b7-6214553bfc8a. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 727.207379] env[61995]: DEBUG oslo_concurrency.lockutils [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] Acquiring lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.207379] env[61995]: DEBUG oslo_concurrency.lockutils [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] Acquired lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.207379] env[61995]: DEBUG nova.network.neutron [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Refreshing network info cache for port 0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.316453] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.391485] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.492020] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 727.606816] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a0b88b-3f39-b15f-b78d-8c2565767d11, 'name': SearchDatastore_Task, 'duration_secs': 0.00937} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.606816] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.606816] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 727.606816] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ab22850-3fd4-467d-aed1-f1ccdade4d56 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.612600] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 727.612600] env[61995]: value = "task-794320" [ 727.612600] env[61995]: _type = "Task" [ 727.612600] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.629020] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794320, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.683198] env[61995]: DEBUG nova.objects.instance [None req-dd7576c3-7cba-4c16-ab91-92b738615eec tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lazy-loading 'pci_requests' on Instance uuid 19fd0358-32e4-434f-b0b3-c8cc6c07ba51 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.686138] env[61995]: DEBUG oslo_vmware.api [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794319, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165032} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.686659] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 727.686844] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 727.687038] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 727.687335] env[61995]: INFO nova.compute.manager [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Took 1.26 seconds to destroy the instance on the hypervisor. [ 727.687437] env[61995]: DEBUG oslo.service.loopingcall [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.687619] env[61995]: DEBUG nova.compute.manager [-] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.687722] env[61995]: DEBUG nova.network.neutron [-] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.895688] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794317, 'name': ReconfigVM_Task, 'duration_secs': 0.882098} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.899026] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Reconfigured VM instance instance-00000022 to attach disk [datastore1] c350194c-ed3e-4996-871a-70b4cb99c7db/c350194c-ed3e-4996-871a-70b4cb99c7db.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 727.900077] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-03602060-ac0f-4439-8334-6e23365925e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.909061] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 727.909061] env[61995]: value = "task-794321" [ 727.909061] env[61995]: _type = "Task" [ 727.909061] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.921330] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794321, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.998822] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 4edbf65d-eaca-47e1-b6bf-03ccb908b52e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 728.125083] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794320, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497441} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.125307] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 728.125533] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.128963] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ac76705-25ff-4370-80b8-ba204437c7ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.129694] env[61995]: DEBUG nova.network.neutron [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updated VIF entry in instance network info cache for port 0f5a4362-281d-470a-95b7-6214553bfc8a. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 728.130065] env[61995]: DEBUG nova.network.neutron [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.132652] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 728.132652] env[61995]: value = "task-794322" [ 728.132652] env[61995]: _type = "Task" [ 728.132652] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.141930] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794322, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.187660] env[61995]: DEBUG nova.objects.base [None req-dd7576c3-7cba-4c16-ab91-92b738615eec tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Object Instance<19fd0358-32e4-434f-b0b3-c8cc6c07ba51> lazy-loaded attributes: flavor,pci_requests {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 728.188475] env[61995]: DEBUG nova.network.neutron [None req-dd7576c3-7cba-4c16-ab91-92b738615eec tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 728.286270] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dd7576c3-7cba-4c16-ab91-92b738615eec tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "interface-19fd0358-32e4-434f-b0b3-c8cc6c07ba51-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.113s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.420151] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794321, 'name': Rename_Task, 'duration_secs': 0.205417} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.420463] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 728.420666] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2aa968a1-590a-4960-be54-78139a77f53d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.426825] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 728.426825] env[61995]: value = "task-794323" [ 728.426825] env[61995]: _type = "Task" [ 728.426825] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.435210] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.501225] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 4da426a1-7af9-4e4e-82ed-c45687a19b86 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 728.508300] env[61995]: DEBUG nova.compute.manager [req-b699e5c6-3e5c-4332-93af-e6685a190f5f req-4ee9af66-2841-4e07-a38a-32dbdeb5ce52 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-vif-deleted-b569b424-48e7-45f4-8f0d-53192e53abf0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.508473] env[61995]: INFO nova.compute.manager [req-b699e5c6-3e5c-4332-93af-e6685a190f5f req-4ee9af66-2841-4e07-a38a-32dbdeb5ce52 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Neutron deleted interface b569b424-48e7-45f4-8f0d-53192e53abf0; detaching it from the instance and deleting it from the info cache [ 728.508743] env[61995]: DEBUG nova.network.neutron [req-b699e5c6-3e5c-4332-93af-e6685a190f5f req-4ee9af66-2841-4e07-a38a-32dbdeb5ce52 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updating instance_info_cache with network_info: [{"id": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "address": "fa:16:3e:96:46:60", "network": {"id": "107ef751-b8eb-47de-914e-e85d23493335", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523281070", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbd71fc-00", "ovs_interfaceid": "ebbd71fc-00dd-4b27-95ed-6670ac9b552e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "address": "fa:16:3e:f5:db:69", "network": {"id": "3a407f9b-a9f0-4839-8dd2-f10ecd904b43", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1652315845", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda35a09f-ea", "ovs_interfaceid": "da35a09f-ea6a-43b8-a817-8a2efbe499bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.636702] env[61995]: DEBUG oslo_concurrency.lockutils [req-346c9b3a-9221-4bf2-8090-195b87589f0d req-ce332b8e-dcab-4db3-9b0e-9f282d46b865 service nova] Releasing lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.645921] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794322, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070377} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.646214] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 728.646974] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae15f06-6c69-4818-849d-1acfab3365a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.669563] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 728.669730] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-253c9557-0598-430e-bc66-3a682cc74401 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.688885] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 728.688885] env[61995]: value = "task-794324" [ 728.688885] env[61995]: _type = "Task" [ 728.688885] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.697401] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794324, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.938733] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794323, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.004889] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 2da17a43-2260-4a82-9ed3-eafd25bb9a6a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 729.011601] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63128660-72ab-4086-b911-aca74721f737 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.021867] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887209f9-01cf-4a15-9584-f2da5047205f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.048834] env[61995]: DEBUG nova.compute.manager [req-b699e5c6-3e5c-4332-93af-e6685a190f5f req-4ee9af66-2841-4e07-a38a-32dbdeb5ce52 service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Detach interface failed, port_id=b569b424-48e7-45f4-8f0d-53192e53abf0, reason: Instance 1c4eb398-c4e9-4588-ba48-2805e3fee750 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 729.077194] env[61995]: DEBUG nova.network.neutron [-] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.198840] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794324, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.439246] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794323, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.507492] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance c51184b6-d2c7-49ff-b061-ab0f9a69cdbf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 729.579572] env[61995]: INFO nova.compute.manager [-] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Took 1.89 seconds to deallocate network for instance. [ 729.643702] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.644660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.698757] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794324, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.938956] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794323, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.011233] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance f6db353f-f7b5-4ba4-8f70-f3b65281da93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 730.086217] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.200227] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794324, 'name': ReconfigVM_Task, 'duration_secs': 1.020622} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.200533] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 45fdae21-bec2-438a-8f2a-c23b0e428057/45fdae21-bec2-438a-8f2a-c23b0e428057.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.201201] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25d9a4d6-3a1a-4196-9cfe-c6857e9a0c88 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.208923] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 730.208923] env[61995]: value = "task-794325" [ 730.208923] env[61995]: _type = "Task" [ 730.208923] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.216939] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794325, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.440400] env[61995]: DEBUG oslo_vmware.api [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794323, 'name': PowerOnVM_Task, 'duration_secs': 1.649526} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.440716] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 730.440963] env[61995]: INFO nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Took 8.41 seconds to spawn the instance on the hypervisor. [ 730.441206] env[61995]: DEBUG nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 730.442082] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a61696-330e-4ceb-abe5-88bb9478bccf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.469729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.469729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.469729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.469729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.469729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.471926] env[61995]: INFO nova.compute.manager [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Terminating instance [ 730.473203] env[61995]: DEBUG nova.compute.manager [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 730.473910] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 730.474788] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba37cc8-0079-4b9a-bd25-c730cc26f86e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.482249] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 730.482808] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d050a4c-f4a0-4fa7-a717-ed3100e61867 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.488686] env[61995]: DEBUG oslo_vmware.api [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 730.488686] env[61995]: value = "task-794326" [ 730.488686] env[61995]: _type = "Task" [ 730.488686] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.497113] env[61995]: DEBUG oslo_vmware.api [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.522021] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance f9c1424f-6096-481c-aa91-acc25bfd8f6e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 730.533774] env[61995]: DEBUG nova.compute.manager [req-ce4cbb93-91a8-42c9-ab0e-3edd340fb7ee req-c3c16db2-45c9-42a9-9429-c5958cea350c service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-vif-deleted-ebbd71fc-00dd-4b27-95ed-6670ac9b552e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.534451] env[61995]: DEBUG nova.compute.manager [req-ce4cbb93-91a8-42c9-ab0e-3edd340fb7ee req-c3c16db2-45c9-42a9-9429-c5958cea350c service nova] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Received event network-vif-deleted-da35a09f-ea6a-43b8-a817-8a2efbe499bb {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.720472] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794325, 'name': Rename_Task, 'duration_secs': 0.166493} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.720741] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 730.720978] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea194252-b340-4215-a356-195ae800005e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.727166] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 730.727166] env[61995]: value = "task-794327" [ 730.727166] env[61995]: _type = "Task" [ 730.727166] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.734524] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794327, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.959978] env[61995]: INFO nova.compute.manager [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Took 34.59 seconds to build instance. [ 730.998796] env[61995]: DEBUG oslo_vmware.api [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794326, 'name': PowerOffVM_Task, 'duration_secs': 0.408075} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.999138] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 730.999334] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 730.999611] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb472265-b0c3-48cf-b065-47835e7ceeb5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.024725] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance eef1cf32-e71e-4696-8170-1023e3a5518d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.064886] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 731.064886] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 731.064886] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Deleting the datastore file [datastore1] 19fd0358-32e4-434f-b0b3-c8cc6c07ba51 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 731.065184] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00344838-67b9-4fdb-82c8-5ce8bb585fd1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.070894] env[61995]: DEBUG oslo_vmware.api [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for the task: (returnval){ [ 731.070894] env[61995]: value = "task-794329" [ 731.070894] env[61995]: _type = "Task" [ 731.070894] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.078317] env[61995]: DEBUG oslo_vmware.api [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.237706] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794327, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.462440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-59e10e5f-9095-486d-9cb2-fc9d9b11dd9f tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.767s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.527566] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d056dadd-492d-4a4d-abc2-fc5f113f865b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.581240] env[61995]: DEBUG oslo_vmware.api [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Task: {'id': task-794329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128928} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.581508] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 731.581689] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 731.581860] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 731.582058] env[61995]: INFO nova.compute.manager [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Took 1.11 seconds to destroy the instance on the hypervisor. [ 731.582315] env[61995]: DEBUG oslo.service.loopingcall [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.582497] env[61995]: DEBUG nova.compute.manager [-] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.582593] env[61995]: DEBUG nova.network.neutron [-] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 731.742319] env[61995]: DEBUG oslo_vmware.api [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794327, 'name': PowerOnVM_Task, 'duration_secs': 0.533346} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.742616] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 731.742815] env[61995]: DEBUG nova.compute.manager [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 731.743806] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca33163-9a94-4727-8bc9-1ef444db39a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.965439] env[61995]: DEBUG nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.030756] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 0d34066f-5b8f-4bac-9b83-67d78987c4b3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 732.031061] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 732.031233] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 732.265608] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.362117] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a1e127-ebe2-4c1b-9f61-14d66fb8e2ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.370711] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841f8367-5c81-48fe-9fd1-fdb9c54f6700 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.400913] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd89959-0cc0-4728-8565-1343b7204406 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.409031] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecbcfdd-93ea-4285-95e7-6f03e3eaadc2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.412741] env[61995]: DEBUG nova.network.neutron [-] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.421875] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 732.484290] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.564039] env[61995]: DEBUG nova.compute.manager [req-25a3bc3f-596a-4e1d-8b1b-7099ca040c86 req-6af44eee-921f-418d-8151-5e6837e5503b service nova] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Received event network-vif-deleted-f97a3a6e-f66d-46ea-9f21-c1f2a63aeea9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.694993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "c350194c-ed3e-4996-871a-70b4cb99c7db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.695307] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.695516] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "c350194c-ed3e-4996-871a-70b4cb99c7db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.695713] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.695863] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.698178] env[61995]: INFO nova.compute.manager [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Terminating instance [ 732.700088] env[61995]: DEBUG nova.compute.manager [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 732.700339] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 732.701179] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649776c0-f098-497f-b2b9-7477b8c61707 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.710661] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 732.710906] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-faba285d-e871-41da-b9bd-efa1727e6da2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.716648] env[61995]: DEBUG oslo_vmware.api [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 732.716648] env[61995]: value = "task-794330" [ 732.716648] env[61995]: _type = "Task" [ 732.716648] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.725162] env[61995]: DEBUG oslo_vmware.api [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.915334] env[61995]: INFO nova.compute.manager [-] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Took 1.33 seconds to deallocate network for instance. [ 732.942504] env[61995]: ERROR nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [req-44aaaae9-d30f-4df2-90cd-251f61463b3f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-44aaaae9-d30f-4df2-90cd-251f61463b3f"}]} [ 732.958701] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 732.973198] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 732.973375] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 732.985035] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 732.998453] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "45fdae21-bec2-438a-8f2a-c23b0e428057" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.999082] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.999195] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "45fdae21-bec2-438a-8f2a-c23b0e428057-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.000258] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.000532] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.003047] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 733.006262] env[61995]: INFO nova.compute.manager [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Terminating instance [ 733.009031] env[61995]: DEBUG nova.compute.manager [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.009031] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 733.012033] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f7c448-f6b0-4e7b-a1b8-ddacde8dcf49 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.017380] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 733.017640] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-468ff9c7-294d-4d71-98db-fd6a6bc5c42d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.023706] env[61995]: DEBUG oslo_vmware.api [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 733.023706] env[61995]: value = "task-794331" [ 733.023706] env[61995]: _type = "Task" [ 733.023706] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.035185] env[61995]: DEBUG oslo_vmware.api [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794331, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.228171] env[61995]: DEBUG oslo_vmware.api [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794330, 'name': PowerOffVM_Task, 'duration_secs': 0.155217} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.228433] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 733.228601] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 733.228849] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d901263-855b-406b-b084-b2f69b7bc95b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.291327] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 733.291327] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 733.291497] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Deleting the datastore file [datastore1] c350194c-ed3e-4996-871a-70b4cb99c7db {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 733.291817] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dccf4b1c-3e26-48c1-bac5-0017dfa297b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.297796] env[61995]: DEBUG oslo_vmware.api [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for the task: (returnval){ [ 733.297796] env[61995]: value = "task-794333" [ 733.297796] env[61995]: _type = "Task" [ 733.297796] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.305018] env[61995]: DEBUG oslo_vmware.api [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794333, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.346620] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ded7766-35f6-4307-9613-8613d2a7a394 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.353999] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4cdf12-7568-451c-bdd0-77bc022514f3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.384720] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e47d48-a8a9-4d99-a008-b52d1811ac93 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.392198] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d18b07-d571-4abb-bd42-b39891631c81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.407193] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 733.422464] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.454426] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.455021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.534163] env[61995]: DEBUG oslo_vmware.api [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794331, 'name': PowerOffVM_Task, 'duration_secs': 0.21902} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.534466] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 733.534595] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 733.534831] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-933f1ab4-0dfa-4a9c-81e2-f0f3d9d2deba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.596355] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 733.596355] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 733.596459] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleting the datastore file [datastore2] 45fdae21-bec2-438a-8f2a-c23b0e428057 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 733.596707] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d08162aa-4451-4e31-8024-a56b7b9306d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.602938] env[61995]: DEBUG oslo_vmware.api [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 733.602938] env[61995]: value = "task-794335" [ 733.602938] env[61995]: _type = "Task" [ 733.602938] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.611299] env[61995]: DEBUG oslo_vmware.api [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794335, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.808314] env[61995]: DEBUG oslo_vmware.api [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Task: {'id': task-794333, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158634} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.808561] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 733.808745] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 733.808928] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 733.809119] env[61995]: INFO nova.compute.manager [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Took 1.11 seconds to destroy the instance on the hypervisor. [ 733.809359] env[61995]: DEBUG oslo.service.loopingcall [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.809539] env[61995]: DEBUG nova.compute.manager [-] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.809632] env[61995]: DEBUG nova.network.neutron [-] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 733.927923] env[61995]: ERROR nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [req-f9c32f5f-03c1-4e18-a08f-6b3ddec597fa] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f9c32f5f-03c1-4e18-a08f-6b3ddec597fa"}]} [ 733.948123] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 733.964137] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 733.964854] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 733.977805] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 734.005531] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 734.111509] env[61995]: DEBUG oslo_vmware.api [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794335, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254959} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.111725] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 734.111908] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 734.112093] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.112267] env[61995]: INFO nova.compute.manager [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Took 1.10 seconds to destroy the instance on the hypervisor. [ 734.112495] env[61995]: DEBUG oslo.service.loopingcall [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.114697] env[61995]: DEBUG nova.compute.manager [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.114799] env[61995]: DEBUG nova.network.neutron [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.385835] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67e0843-38a3-4bf0-a58d-55889d236908 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.394031] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8752883c-b78c-4f0c-99ae-13fa266d6758 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.425300] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a1936a-9d91-4c8d-80ca-a1942766878c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.433035] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ec3ddb-6cff-43b3-9487-cd7232b3cff8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.446579] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 734.567561] env[61995]: DEBUG nova.network.neutron [-] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.648607] env[61995]: DEBUG nova.compute.manager [req-a913b233-9e83-44c8-a5c2-9d55b1ce739a req-7506b951-5807-4fa6-ad20-8f226b044bde service nova] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Received event network-vif-deleted-ef74f12f-3ba8-4392-b066-245f7c8a6425 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.648813] env[61995]: DEBUG nova.compute.manager [req-a913b233-9e83-44c8-a5c2-9d55b1ce739a req-7506b951-5807-4fa6-ad20-8f226b044bde service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Received event network-vif-deleted-515eb152-ff0c-4c68-896f-d46c12f8c848 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.648977] env[61995]: INFO nova.compute.manager [req-a913b233-9e83-44c8-a5c2-9d55b1ce739a req-7506b951-5807-4fa6-ad20-8f226b044bde service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Neutron deleted interface 515eb152-ff0c-4c68-896f-d46c12f8c848; detaching it from the instance and deleting it from the info cache [ 734.649373] env[61995]: DEBUG nova.network.neutron [req-a913b233-9e83-44c8-a5c2-9d55b1ce739a req-7506b951-5807-4fa6-ad20-8f226b044bde service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.978483] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 58 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 734.978725] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 58 to 59 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 734.978875] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 735.001540] env[61995]: DEBUG nova.network.neutron [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.070735] env[61995]: INFO nova.compute.manager [-] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Took 1.26 seconds to deallocate network for instance. [ 735.152706] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dab08ee0-b060-4c3c-9654-4a69d8fee4d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.163023] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf76574-044c-4bef-b3e2-b04e1d5ecd69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.191635] env[61995]: DEBUG nova.compute.manager [req-a913b233-9e83-44c8-a5c2-9d55b1ce739a req-7506b951-5807-4fa6-ad20-8f226b044bde service nova] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Detach interface failed, port_id=515eb152-ff0c-4c68-896f-d46c12f8c848, reason: Instance 45fdae21-bec2-438a-8f2a-c23b0e428057 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 735.485178] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 735.485452] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.071s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.485721] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.370s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.487253] env[61995]: INFO nova.compute.claims [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.504399] env[61995]: INFO nova.compute.manager [-] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Took 1.39 seconds to deallocate network for instance. [ 735.578144] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.010605] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.782612] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11bc221-e378-45ff-8aed-c8fb2905442d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.791256] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbb59eb-af5e-425c-b9ea-ce6378495579 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.819786] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1a6b9e-5525-4d65-9ba7-40161c9532fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.826845] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792bff26-0cd7-4a3e-9b43-a5a1ac1ea970 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.839282] env[61995]: DEBUG nova.compute.provider_tree [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.342157] env[61995]: DEBUG nova.scheduler.client.report [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.846877] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.847451] env[61995]: DEBUG nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.850036] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.797s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.851421] env[61995]: INFO nova.compute.claims [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.356120] env[61995]: DEBUG nova.compute.utils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.360665] env[61995]: DEBUG nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.360755] env[61995]: DEBUG nova.network.neutron [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 738.408715] env[61995]: DEBUG nova.policy [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54b695c82ecc4a95a34d0e67f082f8a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e7ab932afb04a369c1a639b4cba11d0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 738.672820] env[61995]: DEBUG nova.network.neutron [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Successfully created port: 7cc9ac68-fc87-4772-ae4e-041105424b5e {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.863953] env[61995]: DEBUG nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.189572] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6095f8e0-9586-424a-8667-204b223cd60b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.197298] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959f5635-c91f-4743-936f-20225305effc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.226535] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204f3f69-4058-44ea-a492-ee1d75a35ee5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.234041] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6687a0-4495-4d4f-bbf6-8010e29bf54b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.245955] env[61995]: DEBUG nova.compute.provider_tree [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.749427] env[61995]: DEBUG nova.scheduler.client.report [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.873783] env[61995]: DEBUG nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.899995] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.900260] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.900415] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.900592] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.900741] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.900885] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.901098] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.901260] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.901426] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.901587] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.901754] env[61995]: DEBUG nova.virt.hardware [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.902625] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c097fe-551f-4ff1-a03d-8fcd2ac63e5e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.912448] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ebc614-b691-415b-b671-58055596aa62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.153776] env[61995]: DEBUG nova.compute.manager [req-08585cba-75a0-49ab-b155-e90bd59f869a req-623c8413-d0cc-42c7-8510-ed017dfdfe66 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Received event network-vif-plugged-7cc9ac68-fc87-4772-ae4e-041105424b5e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 740.154036] env[61995]: DEBUG oslo_concurrency.lockutils [req-08585cba-75a0-49ab-b155-e90bd59f869a req-623c8413-d0cc-42c7-8510-ed017dfdfe66 service nova] Acquiring lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.154216] env[61995]: DEBUG oslo_concurrency.lockutils [req-08585cba-75a0-49ab-b155-e90bd59f869a req-623c8413-d0cc-42c7-8510-ed017dfdfe66 service nova] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.154384] env[61995]: DEBUG oslo_concurrency.lockutils [req-08585cba-75a0-49ab-b155-e90bd59f869a req-623c8413-d0cc-42c7-8510-ed017dfdfe66 service nova] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.154550] env[61995]: DEBUG nova.compute.manager [req-08585cba-75a0-49ab-b155-e90bd59f869a req-623c8413-d0cc-42c7-8510-ed017dfdfe66 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] No waiting events found dispatching network-vif-plugged-7cc9ac68-fc87-4772-ae4e-041105424b5e {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 740.154712] env[61995]: WARNING nova.compute.manager [req-08585cba-75a0-49ab-b155-e90bd59f869a req-623c8413-d0cc-42c7-8510-ed017dfdfe66 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Received unexpected event network-vif-plugged-7cc9ac68-fc87-4772-ae4e-041105424b5e for instance with vm_state building and task_state spawning. [ 740.255396] env[61995]: DEBUG nova.network.neutron [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Successfully updated port: 7cc9ac68-fc87-4772-ae4e-041105424b5e {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 740.257148] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.257640] env[61995]: DEBUG nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.264840] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.674s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.266628] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.002s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.270936] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.535s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.273118] env[61995]: INFO nova.compute.claims [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.310148] env[61995]: INFO nova.scheduler.client.report [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Deleted allocations for instance e78f8718-e0be-4b2a-ab70-9565ce01ab9d [ 740.766677] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "refresh_cache-a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.766677] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquired lock "refresh_cache-a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.766677] env[61995]: DEBUG nova.network.neutron [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 740.780773] env[61995]: DEBUG nova.compute.utils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.782867] env[61995]: DEBUG nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.783050] env[61995]: DEBUG nova.network.neutron [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 740.823642] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cf30c4d-c9ee-4a92-af65-16f0cebc7184 tempest-ServerDiagnosticsV248Test-267817391 tempest-ServerDiagnosticsV248Test-267817391-project-member] Lock "e78f8718-e0be-4b2a-ab70-9565ce01ab9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.103s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.856430] env[61995]: DEBUG nova.policy [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce97dbfe04a349738a0e5ee9cf177338', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a269227b55044d38154a50fbc722723', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 741.202326] env[61995]: DEBUG nova.network.neutron [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Successfully created port: 22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.283710] env[61995]: DEBUG nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.312843] env[61995]: DEBUG nova.network.neutron [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.477185] env[61995]: DEBUG nova.network.neutron [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Updating instance_info_cache with network_info: [{"id": "7cc9ac68-fc87-4772-ae4e-041105424b5e", "address": "fa:16:3e:3c:81:d6", "network": {"id": "8a5d6007-640d-428c-b538-a5876227271b", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-102919381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e7ab932afb04a369c1a639b4cba11d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cc9ac68-fc", "ovs_interfaceid": "7cc9ac68-fc87-4772-ae4e-041105424b5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.681541] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90acc7b-6068-4586-9234-defd0d5f3512 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.689762] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c9f704-0229-4fea-8699-2411338598db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.723176] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c6eefb-5451-49a2-82c1-d85f60f320c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.730843] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9acf092-4a30-4202-a7f4-d9b507078204 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.743887] env[61995]: DEBUG nova.compute.provider_tree [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.980111] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Releasing lock "refresh_cache-a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.980662] env[61995]: DEBUG nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Instance network_info: |[{"id": "7cc9ac68-fc87-4772-ae4e-041105424b5e", "address": "fa:16:3e:3c:81:d6", "network": {"id": "8a5d6007-640d-428c-b538-a5876227271b", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-102919381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e7ab932afb04a369c1a639b4cba11d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cc9ac68-fc", "ovs_interfaceid": "7cc9ac68-fc87-4772-ae4e-041105424b5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 741.981152] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:81:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1323cb03-8367-485a-962e-131af8eba474', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7cc9ac68-fc87-4772-ae4e-041105424b5e', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 741.988885] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Creating folder: Project (2e7ab932afb04a369c1a639b4cba11d0). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 741.989210] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-653d3a78-f5f9-4d24-ad90-61d381b44c2e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.002285] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Created folder: Project (2e7ab932afb04a369c1a639b4cba11d0) in parent group-v185203. [ 742.002285] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Creating folder: Instances. Parent ref: group-v185247. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.002285] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3442a60-ea61-4de9-8d96-d59528601312 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.009438] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Created folder: Instances in parent group-v185247. [ 742.009678] env[61995]: DEBUG oslo.service.loopingcall [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.009863] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 742.010517] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f01d2a6a-848d-4d4a-bc1e-e8a6f98cae8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.033134] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.033134] env[61995]: value = "task-794338" [ 742.033134] env[61995]: _type = "Task" [ 742.033134] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.041154] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794338, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.180134] env[61995]: DEBUG nova.compute.manager [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Received event network-changed-7cc9ac68-fc87-4772-ae4e-041105424b5e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 742.180411] env[61995]: DEBUG nova.compute.manager [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Refreshing instance network info cache due to event network-changed-7cc9ac68-fc87-4772-ae4e-041105424b5e. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 742.180714] env[61995]: DEBUG oslo_concurrency.lockutils [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] Acquiring lock "refresh_cache-a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.180887] env[61995]: DEBUG oslo_concurrency.lockutils [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] Acquired lock "refresh_cache-a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.181148] env[61995]: DEBUG nova.network.neutron [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Refreshing network info cache for port 7cc9ac68-fc87-4772-ae4e-041105424b5e {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.247207] env[61995]: DEBUG nova.scheduler.client.report [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.295983] env[61995]: DEBUG nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.333937] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.333937] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.333937] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.333937] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.333937] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.334281] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.334281] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.335025] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.335025] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.335025] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.335025] env[61995]: DEBUG nova.virt.hardware [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.335803] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302228d3-9a21-4852-8ce6-593620e9a6a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.344431] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f062dd7e-85a2-477a-ab39-8dfd0d3ab20d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.544071] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794338, 'name': CreateVM_Task, 'duration_secs': 0.326355} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.544268] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 742.544960] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.545145] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.545468] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 742.545706] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4845baf8-0574-49ec-8d0c-f8ca0308a2a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.550273] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 742.550273] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dd9ae8-ec5e-6d09-b4ca-7e448b1c73ff" [ 742.550273] env[61995]: _type = "Task" [ 742.550273] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.557801] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52dd9ae8-ec5e-6d09-b4ca-7e448b1c73ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.749863] env[61995]: DEBUG nova.network.neutron [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Successfully updated port: 22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 742.752349] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.752994] env[61995]: DEBUG nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 742.756128] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.083s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.756390] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.758448] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.077s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.758839] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.760471] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.781s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.761867] env[61995]: INFO nova.compute.claims [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.786283] env[61995]: INFO nova.scheduler.client.report [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleted allocations for instance 08489e01-2ce3-45a1-8f87-2a4c89230aa2 [ 742.787888] env[61995]: INFO nova.scheduler.client.report [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Deleted allocations for instance f47e5bbb-f4a8-4686-8195-8348d2ffb08d [ 743.001958] env[61995]: DEBUG nova.network.neutron [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Updated VIF entry in instance network info cache for port 7cc9ac68-fc87-4772-ae4e-041105424b5e. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.002460] env[61995]: DEBUG nova.network.neutron [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Updating instance_info_cache with network_info: [{"id": "7cc9ac68-fc87-4772-ae4e-041105424b5e", "address": "fa:16:3e:3c:81:d6", "network": {"id": "8a5d6007-640d-428c-b538-a5876227271b", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-102919381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e7ab932afb04a369c1a639b4cba11d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cc9ac68-fc", "ovs_interfaceid": "7cc9ac68-fc87-4772-ae4e-041105424b5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.063762] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52dd9ae8-ec5e-6d09-b4ca-7e448b1c73ff, 'name': SearchDatastore_Task, 'duration_secs': 0.009706} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.063762] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.063762] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 743.063762] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.064127] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.064218] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.064461] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-936ada24-49fa-4650-be7a-c2de91a063db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.072942] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.073953] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 743.073953] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa4d36ff-951d-4f97-b025-e943e256fb70 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.079728] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 743.079728] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5253ab9c-b134-e87e-d376-fec3ca230164" [ 743.079728] env[61995]: _type = "Task" [ 743.079728] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.087084] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5253ab9c-b134-e87e-d376-fec3ca230164, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.253131] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "refresh_cache-c4f744f9-36f3-4913-9ced-f0db93d3b4df" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.253401] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "refresh_cache-c4f744f9-36f3-4913-9ced-f0db93d3b4df" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.253471] env[61995]: DEBUG nova.network.neutron [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.268359] env[61995]: DEBUG nova.compute.utils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.270516] env[61995]: DEBUG nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 743.270692] env[61995]: DEBUG nova.network.neutron [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 743.301204] env[61995]: DEBUG oslo_concurrency.lockutils [None req-874edf07-48d6-4020-9b54-2f6a63eae9c6 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "08489e01-2ce3-45a1-8f87-2a4c89230aa2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.157s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.322756] env[61995]: DEBUG oslo_concurrency.lockutils [None req-376fc3de-2f40-42d3-a6b4-1a5921b40356 tempest-MultipleCreateTestJSON-612890301 tempest-MultipleCreateTestJSON-612890301-project-member] Lock "f47e5bbb-f4a8-4686-8195-8348d2ffb08d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.408s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.355216] env[61995]: DEBUG nova.policy [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce97dbfe04a349738a0e5ee9cf177338', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a269227b55044d38154a50fbc722723', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 743.505242] env[61995]: DEBUG oslo_concurrency.lockutils [req-a159e932-f189-47a2-993d-b3520a7f4952 req-ab9c5763-7636-491e-a885-2fd1065b6813 service nova] Releasing lock "refresh_cache-a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.592414] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5253ab9c-b134-e87e-d376-fec3ca230164, 'name': SearchDatastore_Task, 'duration_secs': 0.00876} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.593050] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9954c46a-dfb8-43cc-a3f2-58fc131dff85 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.599184] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 743.599184] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52bcbe6b-c5ef-b977-5dcc-0edda5041088" [ 743.599184] env[61995]: _type = "Task" [ 743.599184] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.608017] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52bcbe6b-c5ef-b977-5dcc-0edda5041088, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.774441] env[61995]: DEBUG nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.807832] env[61995]: DEBUG nova.network.neutron [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.838683] env[61995]: DEBUG nova.network.neutron [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Successfully created port: 654127d3-17e0-4939-b4b5-6d2e905f4f35 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.072599] env[61995]: DEBUG nova.network.neutron [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Updating instance_info_cache with network_info: [{"id": "22b10d97-e8c2-4f19-b6b6-d5a55fe432c8", "address": "fa:16:3e:ec:68:10", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b10d97-e8", "ovs_interfaceid": "22b10d97-e8c2-4f19-b6b6-d5a55fe432c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.113295] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52bcbe6b-c5ef-b977-5dcc-0edda5041088, 'name': SearchDatastore_Task, 'duration_secs': 0.009124} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.113295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.113295] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] a04fbe7e-e53e-43ef-9a6f-ba3895ce7597/a04fbe7e-e53e-43ef-9a6f-ba3895ce7597.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 744.113623] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c5ea953-715e-4171-b626-26034acfa5e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.121561] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 744.121561] env[61995]: value = "task-794339" [ 744.121561] env[61995]: _type = "Task" [ 744.121561] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.122982] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c39004b-32f6-40b7-8a75-a4f327c0609d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.135546] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a829b8a3-914e-4d1a-95dc-2835cb8d21bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.139294] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.171964] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08dfd100-c74c-4345-abb6-63e8b0127132 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.179545] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c867be7-8132-4fa8-bf4e-270a6e3787a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.192854] env[61995]: DEBUG nova.compute.provider_tree [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.267962] env[61995]: DEBUG nova.compute.manager [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Received event network-vif-plugged-22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.268297] env[61995]: DEBUG oslo_concurrency.lockutils [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] Acquiring lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.268481] env[61995]: DEBUG oslo_concurrency.lockutils [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.268642] env[61995]: DEBUG oslo_concurrency.lockutils [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.268806] env[61995]: DEBUG nova.compute.manager [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] No waiting events found dispatching network-vif-plugged-22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 744.269076] env[61995]: WARNING nova.compute.manager [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Received unexpected event network-vif-plugged-22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 for instance with vm_state building and task_state spawning. [ 744.269277] env[61995]: DEBUG nova.compute.manager [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Received event network-changed-22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.269430] env[61995]: DEBUG nova.compute.manager [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Refreshing instance network info cache due to event network-changed-22b10d97-e8c2-4f19-b6b6-d5a55fe432c8. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 744.269597] env[61995]: DEBUG oslo_concurrency.lockutils [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] Acquiring lock "refresh_cache-c4f744f9-36f3-4913-9ced-f0db93d3b4df" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.579577] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "refresh_cache-c4f744f9-36f3-4913-9ced-f0db93d3b4df" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.580522] env[61995]: DEBUG nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance network_info: |[{"id": "22b10d97-e8c2-4f19-b6b6-d5a55fe432c8", "address": "fa:16:3e:ec:68:10", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b10d97-e8", "ovs_interfaceid": "22b10d97-e8c2-4f19-b6b6-d5a55fe432c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 744.580522] env[61995]: DEBUG oslo_concurrency.lockutils [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] Acquired lock "refresh_cache-c4f744f9-36f3-4913-9ced-f0db93d3b4df" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.580522] env[61995]: DEBUG nova.network.neutron [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Refreshing network info cache for port 22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.584288] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:68:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22b10d97-e8c2-4f19-b6b6-d5a55fe432c8', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.589839] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating folder: Project (0a269227b55044d38154a50fbc722723). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 744.590653] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-459902d9-1043-40bc-9da5-31e2657d8586 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.601653] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created folder: Project (0a269227b55044d38154a50fbc722723) in parent group-v185203. [ 744.601859] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating folder: Instances. Parent ref: group-v185250. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 744.602140] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a07b93d-3466-400f-ae4b-881dd7eaac06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.611256] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created folder: Instances in parent group-v185250. [ 744.611769] env[61995]: DEBUG oslo.service.loopingcall [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.611769] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 744.611870] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6671d39-8b09-42d7-81fb-8e9248711e88 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.636851] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.423429} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.638189] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] a04fbe7e-e53e-43ef-9a6f-ba3895ce7597/a04fbe7e-e53e-43ef-9a6f-ba3895ce7597.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 744.638533] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.638781] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.638781] env[61995]: value = "task-794342" [ 744.638781] env[61995]: _type = "Task" [ 744.638781] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.638942] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6dd1b389-a662-4259-9fb2-111192a373e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.649833] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794342, 'name': CreateVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.651112] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 744.651112] env[61995]: value = "task-794343" [ 744.651112] env[61995]: _type = "Task" [ 744.651112] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.659290] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794343, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.695942] env[61995]: DEBUG nova.scheduler.client.report [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.789739] env[61995]: DEBUG nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 744.824899] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.824899] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.824899] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.824899] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.824899] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.824899] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.831482] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.831482] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.831482] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.831482] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.831482] env[61995]: DEBUG nova.virt.hardware [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.831482] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9a4d06-40c5-448c-b833-84114feb0470 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.842747] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2daf64f-d1dd-48b8-b5dc-1b43dae771c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.150510] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794342, 'name': CreateVM_Task, 'duration_secs': 0.440685} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.150580] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 745.151297] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.151455] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.151762] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 745.154813] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49d7f316-8861-4a76-899e-079b0f3fca05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.164159] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 745.164159] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527f8b32-6753-084b-cf61-9fc4898ff65c" [ 745.164159] env[61995]: _type = "Task" [ 745.164159] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.166980] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794343, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074166} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.170270] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.170960] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517fd72b-6681-4d31-82eb-e0bd6ba8b9c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.184916] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527f8b32-6753-084b-cf61-9fc4898ff65c, 'name': SearchDatastore_Task, 'duration_secs': 0.009363} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.194816] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.195106] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.197023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.197023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.197023] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.208288] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] a04fbe7e-e53e-43ef-9a6f-ba3895ce7597/a04fbe7e-e53e-43ef-9a6f-ba3895ce7597.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.208288] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4d892ef-0f80-451f-ba13-c75c3bed4b0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.209781] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.211655] env[61995]: DEBUG nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.214139] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b401ee0f-168c-41b7-9103-4a44eaf10894 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.232881] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.156s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.234468] env[61995]: INFO nova.compute.claims [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.245914] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 745.245914] env[61995]: value = "task-794344" [ 745.245914] env[61995]: _type = "Task" [ 745.245914] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.247117] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.247298] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 745.251768] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42538597-2d34-4f9f-915d-57df4e859075 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.259021] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794344, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.261963] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 745.261963] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c258d9-d046-7ad0-fb4d-749efdd6a7e1" [ 745.261963] env[61995]: _type = "Task" [ 745.261963] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.268273] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c258d9-d046-7ad0-fb4d-749efdd6a7e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.482614] env[61995]: DEBUG nova.network.neutron [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Updated VIF entry in instance network info cache for port 22b10d97-e8c2-4f19-b6b6-d5a55fe432c8. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 745.482985] env[61995]: DEBUG nova.network.neutron [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Updating instance_info_cache with network_info: [{"id": "22b10d97-e8c2-4f19-b6b6-d5a55fe432c8", "address": "fa:16:3e:ec:68:10", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b10d97-e8", "ovs_interfaceid": "22b10d97-e8c2-4f19-b6b6-d5a55fe432c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.735220] env[61995]: DEBUG nova.compute.utils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.736701] env[61995]: DEBUG nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Not allocating networking since 'none' was specified. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 745.757374] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794344, 'name': ReconfigVM_Task, 'duration_secs': 0.277553} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.757704] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Reconfigured VM instance instance-00000023 to attach disk [datastore1] a04fbe7e-e53e-43ef-9a6f-ba3895ce7597/a04fbe7e-e53e-43ef-9a6f-ba3895ce7597.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.758394] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-479fe71e-cdb0-4d06-bf76-83cf3232f64c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.765461] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 745.765461] env[61995]: value = "task-794345" [ 745.765461] env[61995]: _type = "Task" [ 745.765461] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.772975] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c258d9-d046-7ad0-fb4d-749efdd6a7e1, 'name': SearchDatastore_Task, 'duration_secs': 0.011895} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.773972] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39f909f8-4803-471c-ad6f-894ef6c97448 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.782201] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794345, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.789801] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 745.789801] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a89eed-7be6-d468-1b82-1551db858a4c" [ 745.789801] env[61995]: _type = "Task" [ 745.789801] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.803657] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a89eed-7be6-d468-1b82-1551db858a4c, 'name': SearchDatastore_Task, 'duration_secs': 0.013624} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.803657] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.805120] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 745.805120] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0d86b3d-566f-4e38-ac39-cfe08459609d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.810883] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 745.810883] env[61995]: value = "task-794346" [ 745.810883] env[61995]: _type = "Task" [ 745.810883] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.822039] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794346, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.985397] env[61995]: DEBUG oslo_concurrency.lockutils [req-b75c28d9-a9e0-4e91-9da5-4785c12b2745 req-2eec912c-f208-48aa-a9b9-0d58a56a0b43 service nova] Releasing lock "refresh_cache-c4f744f9-36f3-4913-9ced-f0db93d3b4df" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.079078] env[61995]: DEBUG nova.compute.manager [req-cf87e459-44a7-4e09-a809-c8cb16ff08ea req-2a63d21e-9d1d-42cd-9434-4ad0b7f31c32 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Received event network-vif-plugged-654127d3-17e0-4939-b4b5-6d2e905f4f35 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 746.079250] env[61995]: DEBUG oslo_concurrency.lockutils [req-cf87e459-44a7-4e09-a809-c8cb16ff08ea req-2a63d21e-9d1d-42cd-9434-4ad0b7f31c32 service nova] Acquiring lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.079466] env[61995]: DEBUG oslo_concurrency.lockutils [req-cf87e459-44a7-4e09-a809-c8cb16ff08ea req-2a63d21e-9d1d-42cd-9434-4ad0b7f31c32 service nova] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.079633] env[61995]: DEBUG oslo_concurrency.lockutils [req-cf87e459-44a7-4e09-a809-c8cb16ff08ea req-2a63d21e-9d1d-42cd-9434-4ad0b7f31c32 service nova] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.079801] env[61995]: DEBUG nova.compute.manager [req-cf87e459-44a7-4e09-a809-c8cb16ff08ea req-2a63d21e-9d1d-42cd-9434-4ad0b7f31c32 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] No waiting events found dispatching network-vif-plugged-654127d3-17e0-4939-b4b5-6d2e905f4f35 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 746.079966] env[61995]: WARNING nova.compute.manager [req-cf87e459-44a7-4e09-a809-c8cb16ff08ea req-2a63d21e-9d1d-42cd-9434-4ad0b7f31c32 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Received unexpected event network-vif-plugged-654127d3-17e0-4939-b4b5-6d2e905f4f35 for instance with vm_state building and task_state spawning. [ 746.238906] env[61995]: DEBUG nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.281856] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794345, 'name': Rename_Task, 'duration_secs': 0.142886} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.282165] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 746.282405] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-482a90ca-9757-46ad-bda5-9fbd9d809067 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.296018] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 746.296018] env[61995]: value = "task-794347" [ 746.296018] env[61995]: _type = "Task" [ 746.296018] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.303559] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.323329] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794346, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471398} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.326357] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 746.326357] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.326357] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2c1a4d7-9efc-42f6-a31a-510850d8282b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.332126] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 746.332126] env[61995]: value = "task-794348" [ 746.332126] env[61995]: _type = "Task" [ 746.332126] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.345981] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794348, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.361912] env[61995]: DEBUG nova.network.neutron [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Successfully updated port: 654127d3-17e0-4939-b4b5-6d2e905f4f35 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 746.616899] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40acdcc2-5a7a-420f-a0bc-979c04d60c66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.624497] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a93d8c-12da-4433-af11-d1954d77d70c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.656035] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1206dab8-f882-4905-8972-c44c7d373a77 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.662776] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10612d30-9f06-401b-9f02-c5f198a9a2ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.675889] env[61995]: DEBUG nova.compute.provider_tree [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 746.804460] env[61995]: DEBUG oslo_vmware.api [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794347, 'name': PowerOnVM_Task, 'duration_secs': 0.458116} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.804723] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 746.804910] env[61995]: INFO nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Took 6.93 seconds to spawn the instance on the hypervisor. [ 746.805094] env[61995]: DEBUG nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 746.805851] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96586130-ae87-4193-81de-cd06fbe3d45a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.842866] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794348, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101707} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.842866] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.843196] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b20d8d-e103-4f8c-a5bb-439a1e28f094 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.866735] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.866735] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "refresh_cache-fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.866735] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "refresh_cache-fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.866735] env[61995]: DEBUG nova.network.neutron [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 746.868030] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82b909e7-e8be-49eb-a511-64e9038fb759 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.890841] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 746.890841] env[61995]: value = "task-794349" [ 746.890841] env[61995]: _type = "Task" [ 746.890841] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.899010] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794349, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.931895] env[61995]: DEBUG nova.network.neutron [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.106602] env[61995]: DEBUG nova.network.neutron [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Updating instance_info_cache with network_info: [{"id": "654127d3-17e0-4939-b4b5-6d2e905f4f35", "address": "fa:16:3e:4e:ad:02", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap654127d3-17", "ovs_interfaceid": "654127d3-17e0-4939-b4b5-6d2e905f4f35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.239931] env[61995]: DEBUG nova.scheduler.client.report [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 59 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 747.240250] env[61995]: DEBUG nova.compute.provider_tree [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 59 to 60 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 747.240502] env[61995]: DEBUG nova.compute.provider_tree [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 747.256316] env[61995]: DEBUG nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.315406] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.315743] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.315786] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.316228] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.316228] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.317404] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.317404] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.317404] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.317404] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.317404] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.317404] env[61995]: DEBUG nova.virt.hardware [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.328707] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e064410-65cc-44fb-9931-b8222a7ef447 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.338680] env[61995]: INFO nova.compute.manager [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Took 41.25 seconds to build instance. [ 747.355689] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1cdc6a3-1ba0-4952-800a-f9ba5f536df8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.381568] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 747.388148] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Creating folder: Project (fc5721d360894580a6e172da0cbe6aaa). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 747.388529] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47ade97e-4203-45dc-a590-e4f38ea0a276 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.400067] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794349, 'name': ReconfigVM_Task, 'duration_secs': 0.295263} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.402027] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Reconfigured VM instance instance-00000024 to attach disk [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.402627] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Created folder: Project (fc5721d360894580a6e172da0cbe6aaa) in parent group-v185203. [ 747.402802] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Creating folder: Instances. Parent ref: group-v185253. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 747.403024] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f9de75b-58e1-42c6-8633-30e91d36dae9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.404775] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e0a9eea-e516-4a3f-b038-3885bf7cb04f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.411495] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 747.411495] env[61995]: value = "task-794351" [ 747.411495] env[61995]: _type = "Task" [ 747.411495] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.416070] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Created folder: Instances in parent group-v185253. [ 747.416274] env[61995]: DEBUG oslo.service.loopingcall [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.416451] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 747.416660] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ec1a6b75-6ce0-4a6c-a9b9-ef485ed52e28 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.433439] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794351, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.437754] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 747.437754] env[61995]: value = "task-794353" [ 747.437754] env[61995]: _type = "Task" [ 747.437754] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.445325] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794353, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.611150] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "refresh_cache-fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.611211] env[61995]: DEBUG nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Instance network_info: |[{"id": "654127d3-17e0-4939-b4b5-6d2e905f4f35", "address": "fa:16:3e:4e:ad:02", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap654127d3-17", "ovs_interfaceid": "654127d3-17e0-4939-b4b5-6d2e905f4f35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 747.612342] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:ad:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '654127d3-17e0-4939-b4b5-6d2e905f4f35', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 747.624906] env[61995]: DEBUG oslo.service.loopingcall [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.626217] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 747.627148] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1db98327-57da-4454-980a-7408016768cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.648433] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 747.648433] env[61995]: value = "task-794354" [ 747.648433] env[61995]: _type = "Task" [ 747.648433] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.656443] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794354, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.729829] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.730155] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.751543] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.752033] env[61995]: DEBUG nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.755478] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.629s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.756977] env[61995]: INFO nova.compute.claims [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.847312] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d68920c5-f38a-4203-bb0c-393facb6ee11 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.744s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.924601] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794351, 'name': Rename_Task, 'duration_secs': 0.180212} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.924857] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 747.925490] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56b77bed-752c-4302-822d-ddb39420611a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.932169] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 747.932169] env[61995]: value = "task-794355" [ 747.932169] env[61995]: _type = "Task" [ 747.932169] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.940456] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794355, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.948544] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794353, 'name': CreateVM_Task, 'duration_secs': 0.345785} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.948544] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 747.948779] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.948961] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.949299] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 747.949658] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66c82507-f4cc-4cba-b1fa-9dd6910d7e27 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.956142] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 747.956142] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d0078d-ed31-4400-28e8-8751f6237c49" [ 747.956142] env[61995]: _type = "Task" [ 747.956142] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.964274] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d0078d-ed31-4400-28e8-8751f6237c49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.115901] env[61995]: DEBUG nova.compute.manager [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Received event network-changed-654127d3-17e0-4939-b4b5-6d2e905f4f35 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 748.116154] env[61995]: DEBUG nova.compute.manager [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Refreshing instance network info cache due to event network-changed-654127d3-17e0-4939-b4b5-6d2e905f4f35. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 748.116438] env[61995]: DEBUG oslo_concurrency.lockutils [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] Acquiring lock "refresh_cache-fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.117784] env[61995]: DEBUG oslo_concurrency.lockutils [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] Acquired lock "refresh_cache-fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.118059] env[61995]: DEBUG nova.network.neutron [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Refreshing network info cache for port 654127d3-17e0-4939-b4b5-6d2e905f4f35 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 748.158717] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794354, 'name': CreateVM_Task, 'duration_secs': 0.368211} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.158952] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 748.159561] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.257033] env[61995]: DEBUG nova.compute.utils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.258476] env[61995]: DEBUG nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Not allocating networking since 'none' was specified. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 748.350326] env[61995]: DEBUG nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.443028] env[61995]: DEBUG oslo_vmware.api [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794355, 'name': PowerOnVM_Task, 'duration_secs': 0.480188} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.443398] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 748.443628] env[61995]: INFO nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Took 6.15 seconds to spawn the instance on the hypervisor. [ 748.443850] env[61995]: DEBUG nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 748.444668] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943e539a-cc6e-4c84-899d-792561a05705 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.465036] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d0078d-ed31-4400-28e8-8751f6237c49, 'name': SearchDatastore_Task, 'duration_secs': 0.013475} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.465036] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.465036] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.465236] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.465293] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.465480] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 748.465911] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.468474] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 748.468474] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5691117-ba64-4ed4-b97a-73ec177c98b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.469911] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0be71989-370f-4a46-9289-4b38df45b413 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.475367] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 748.475367] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e6f719-33cb-7f24-3ea5-bcef6bf8fa52" [ 748.475367] env[61995]: _type = "Task" [ 748.475367] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.480283] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 748.480461] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 748.481488] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b8ab721-4a17-4d65-b195-49171cb4baa2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.487888] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e6f719-33cb-7f24-3ea5-bcef6bf8fa52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.490300] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 748.490300] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c19f4a-a207-7a02-2aad-f7642af1e4b3" [ 748.490300] env[61995]: _type = "Task" [ 748.490300] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.499725] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c19f4a-a207-7a02-2aad-f7642af1e4b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.760462] env[61995]: DEBUG nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.886075] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.969725] env[61995]: INFO nova.compute.manager [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Took 38.95 seconds to build instance. [ 748.981203] env[61995]: DEBUG nova.network.neutron [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Updated VIF entry in instance network info cache for port 654127d3-17e0-4939-b4b5-6d2e905f4f35. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 748.981768] env[61995]: DEBUG nova.network.neutron [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Updating instance_info_cache with network_info: [{"id": "654127d3-17e0-4939-b4b5-6d2e905f4f35", "address": "fa:16:3e:4e:ad:02", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap654127d3-17", "ovs_interfaceid": "654127d3-17e0-4939-b4b5-6d2e905f4f35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.986982] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e6f719-33cb-7f24-3ea5-bcef6bf8fa52, 'name': SearchDatastore_Task, 'duration_secs': 0.016432} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.990896] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.992011] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.992011] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.002092] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c19f4a-a207-7a02-2aad-f7642af1e4b3, 'name': SearchDatastore_Task, 'duration_secs': 0.041605} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.006102] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-824aaf78-9ab5-49bc-b71b-5f21954310ce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.013742] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 749.013742] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]524db2f4-ebc4-e323-298f-c90d78ee17e7" [ 749.013742] env[61995]: _type = "Task" [ 749.013742] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.024458] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524db2f4-ebc4-e323-298f-c90d78ee17e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.127187] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07f320c-c3bd-48bd-af01-5db0334baac7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.135539] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0882b77-d907-479f-b91c-d4b6163c12e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.176230] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023b3062-fe63-4be6-8b81-b7f9d5e4bc55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.187017] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884b9bd2-1aca-4547-be07-6a61796035a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.198203] env[61995]: DEBUG nova.compute.provider_tree [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.471104] env[61995]: DEBUG oslo_concurrency.lockutils [None req-762be958-ebf6-4c3d-addc-75c4446cc665 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.037s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.491909] env[61995]: DEBUG oslo_concurrency.lockutils [req-a2be6b7b-2e35-42a6-a568-86f199efb74f req-d9ffd076-0a9e-4d2d-bf8a-73d6ace0e956 service nova] Releasing lock "refresh_cache-fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.527913] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524db2f4-ebc4-e323-298f-c90d78ee17e7, 'name': SearchDatastore_Task, 'duration_secs': 0.026618} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.528369] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.528648] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 55555deb-9883-483b-a901-42b1a66c60d6/55555deb-9883-483b-a901-42b1a66c60d6.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 749.528937] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.529173] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 749.529384] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b878d9f4-d89b-4395-bd69-a017c5d54eec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.531629] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fde11ead-6cf0-41e4-a488-69dfa61c267e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.540135] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 749.540135] env[61995]: value = "task-794356" [ 749.540135] env[61995]: _type = "Task" [ 749.540135] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.544423] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 749.544682] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 749.545768] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c056e8c-0ca0-45b2-baeb-a8ec7944d087 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.551426] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794356, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.555221] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 749.555221] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f266b1-2d04-e371-c865-7415b7abc2bb" [ 749.555221] env[61995]: _type = "Task" [ 749.555221] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.563604] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f266b1-2d04-e371-c865-7415b7abc2bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.701669] env[61995]: DEBUG nova.scheduler.client.report [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.775563] env[61995]: DEBUG nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.806828] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.807226] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.807450] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.807842] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.809377] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.809601] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.809892] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.810128] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.810365] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.811244] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.811900] env[61995]: DEBUG nova.virt.hardware [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.813338] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e7ebb2-3079-438b-94cd-1d404fa8c0a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.825557] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0425554-b498-47db-97cf-dfce5a981ac2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.842255] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.848252] env[61995]: DEBUG oslo.service.loopingcall [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.848625] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 749.848999] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-398b408f-da32-455f-8eb8-7d1be71e0193 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.865717] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.865717] env[61995]: value = "task-794357" [ 749.865717] env[61995]: _type = "Task" [ 749.865717] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.876029] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794357, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.976447] env[61995]: DEBUG nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.052660] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794356, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.068485] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f266b1-2d04-e371-c865-7415b7abc2bb, 'name': SearchDatastore_Task, 'duration_secs': 0.00983} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.068912] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fff960d-0f98-4d0e-99d4-6ccb67eaf435 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.077114] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 750.077114] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52efe616-eb7c-3852-a4f6-7540e22886d1" [ 750.077114] env[61995]: _type = "Task" [ 750.077114] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.091150] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52efe616-eb7c-3852-a4f6-7540e22886d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.207361] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.208034] env[61995]: DEBUG nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.211034] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.560s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.211157] env[61995]: DEBUG nova.objects.instance [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lazy-loading 'resources' on Instance uuid 24db5fad-52fd-4689-a8aa-2ae4b0a06710 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.376669] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794357, 'name': CreateVM_Task, 'duration_secs': 0.388499} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.376775] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 750.382598] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.382598] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.382598] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 750.382598] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27c41a75-ac62-42f8-9f54-d5c8a3c7db5c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.382990] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 750.382990] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529af7e3-71ab-b433-bd1b-374a18f99925" [ 750.382990] env[61995]: _type = "Task" [ 750.382990] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.390754] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529af7e3-71ab-b433-bd1b-374a18f99925, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.504480] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.551576] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794356, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762736} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.551848] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 55555deb-9883-483b-a901-42b1a66c60d6/55555deb-9883-483b-a901-42b1a66c60d6.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 750.552156] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.552922] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9d485bd-818c-41c6-8428-a0706c1bad06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.559368] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 750.559368] env[61995]: value = "task-794358" [ 750.559368] env[61995]: _type = "Task" [ 750.559368] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.569756] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794358, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.586747] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52efe616-eb7c-3852-a4f6-7540e22886d1, 'name': SearchDatastore_Task, 'duration_secs': 0.053564} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.587040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.587327] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] fcdc7c33-8c46-4f4d-bb39-e6f943256cb0/fcdc7c33-8c46-4f4d-bb39-e6f943256cb0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 750.587604] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c601900-27dc-458e-be0b-d42e9f59b413 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.593970] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 750.593970] env[61995]: value = "task-794359" [ 750.593970] env[61995]: _type = "Task" [ 750.593970] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.601962] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794359, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.717276] env[61995]: DEBUG nova.compute.utils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.728215] env[61995]: DEBUG nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 750.728215] env[61995]: DEBUG nova.network.neutron [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 750.784114] env[61995]: DEBUG nova.policy [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62c2af6ac5304f01b084b89d2a67d1ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ec923f4df4f44d19fd832b77dfda303', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 750.895092] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529af7e3-71ab-b433-bd1b-374a18f99925, 'name': SearchDatastore_Task, 'duration_secs': 0.032468} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.898214] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.898519] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.898787] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.899009] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.899254] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.899760] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7a28320-e573-4545-ac68-a482fa84c6b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.908103] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 750.908241] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 750.908949] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f8b7f66-4d97-4aa4-892a-f752c265a8bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.919128] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 750.919128] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522061b7-63d2-1fd1-0ed1-0be2635666f5" [ 750.919128] env[61995]: _type = "Task" [ 750.919128] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.927890] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522061b7-63d2-1fd1-0ed1-0be2635666f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.069093] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794358, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073288} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.072097] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.073381] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b5d1ad-4fbd-4a4e-b071-5cda49cbfb47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.097082] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 55555deb-9883-483b-a901-42b1a66c60d6/55555deb-9883-483b-a901-42b1a66c60d6.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.101744] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41c1c1c8-daa9-447b-b74e-13595fdca16a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.139231] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794359, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.140455] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 751.140455] env[61995]: value = "task-794360" [ 751.140455] env[61995]: _type = "Task" [ 751.140455] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.153976] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794360, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.181775] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8db411-5fe4-434f-87ef-bba26fcb168d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.193238] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f7b5c4-b647-4559-a6c5-98478e7fac86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.228150] env[61995]: DEBUG nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.233304] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535d20a7-bcf9-4a5e-b65c-5c10950d6957 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.237408] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.238993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.238993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.238993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.238993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.241983] env[61995]: INFO nova.compute.manager [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Terminating instance [ 751.247511] env[61995]: DEBUG nova.compute.manager [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.247903] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.249582] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cb4155-8ea5-423b-b0b4-3e8cbe29e0fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.256610] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc26bd8-ef13-4f40-84a7-ba739fff897d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.271486] env[61995]: DEBUG nova.compute.provider_tree [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.273597] env[61995]: DEBUG nova.network.neutron [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Successfully created port: a1cabe4c-61ba-460d-b384-421305897add {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.277806] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 751.278645] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a05fd780-979a-41b5-98e6-5605757e19bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.284883] env[61995]: DEBUG oslo_vmware.api [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 751.284883] env[61995]: value = "task-794361" [ 751.284883] env[61995]: _type = "Task" [ 751.284883] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.295183] env[61995]: DEBUG oslo_vmware.api [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.429048] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522061b7-63d2-1fd1-0ed1-0be2635666f5, 'name': SearchDatastore_Task, 'duration_secs': 0.011462} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.429857] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c7f98c3-fe8c-46c4-be90-5eb34ca6d545 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.435940] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 751.435940] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52014c80-f6cc-5918-b9f6-e08830c5f0f3" [ 751.435940] env[61995]: _type = "Task" [ 751.435940] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.445435] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52014c80-f6cc-5918-b9f6-e08830c5f0f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.612236] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794359, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681726} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.612659] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] fcdc7c33-8c46-4f4d-bb39-e6f943256cb0/fcdc7c33-8c46-4f4d-bb39-e6f943256cb0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 751.612879] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.612987] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b8eb225-e36b-45d6-afe9-d343a085964a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.621234] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 751.621234] env[61995]: value = "task-794362" [ 751.621234] env[61995]: _type = "Task" [ 751.621234] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.628840] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794362, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.653416] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794360, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.783039] env[61995]: DEBUG nova.scheduler.client.report [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.806206] env[61995]: DEBUG oslo_vmware.api [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794361, 'name': PowerOffVM_Task, 'duration_secs': 0.218996} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.806738] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 751.806999] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 751.809157] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d06d57ea-5011-4102-ac18-e8fded8c70de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.877822] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 751.877917] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 751.878217] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Deleting the datastore file [datastore1] a04fbe7e-e53e-43ef-9a6f-ba3895ce7597 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.878470] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd1b9fde-4a69-4586-bd45-afe4603c64a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.884021] env[61995]: DEBUG oslo_vmware.api [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for the task: (returnval){ [ 751.884021] env[61995]: value = "task-794364" [ 751.884021] env[61995]: _type = "Task" [ 751.884021] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.892541] env[61995]: DEBUG oslo_vmware.api [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.952938] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52014c80-f6cc-5918-b9f6-e08830c5f0f3, 'name': SearchDatastore_Task, 'duration_secs': 0.01228} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.953310] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.953630] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 751.953925] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5a476ab-c427-4b64-9cfa-cb9530abb766 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.966470] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 751.966470] env[61995]: value = "task-794365" [ 751.966470] env[61995]: _type = "Task" [ 751.966470] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.976972] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.041486] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "b40b1866-ced3-40ef-9ab7-30d58b75f288" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.041760] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.130270] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794362, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07023} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.130593] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.131450] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76004135-f57e-4205-b5f7-a1efa4a48b7f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.153568] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] fcdc7c33-8c46-4f4d-bb39-e6f943256cb0/fcdc7c33-8c46-4f4d-bb39-e6f943256cb0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.156911] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96e1eae8-b74d-42b0-ab71-b6f2583a6621 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.176715] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794360, 'name': ReconfigVM_Task, 'duration_secs': 0.712315} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.176995] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 55555deb-9883-483b-a901-42b1a66c60d6/55555deb-9883-483b-a901-42b1a66c60d6.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.177634] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a549e239-e9d2-4d56-8a6a-d415efb8190c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.180548] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 752.180548] env[61995]: value = "task-794366" [ 752.180548] env[61995]: _type = "Task" [ 752.180548] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.187350] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 752.187350] env[61995]: value = "task-794367" [ 752.187350] env[61995]: _type = "Task" [ 752.187350] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.191045] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.199120] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794367, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.243971] env[61995]: DEBUG nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.272401] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.272670] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.273404] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.273404] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.273404] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.273581] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.273694] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.273901] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.274094] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.274376] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.274542] env[61995]: DEBUG nova.virt.hardware [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.275599] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a882c9d6-0ecb-4b24-ac2b-8841955d5f4d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.285955] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dba747f-23bf-4b94-8c42-a1866fa5ed27 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.301110] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.090s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.304504] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.706s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.306286] env[61995]: INFO nova.compute.claims [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.333683] env[61995]: INFO nova.scheduler.client.report [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Deleted allocations for instance 24db5fad-52fd-4689-a8aa-2ae4b0a06710 [ 752.396819] env[61995]: DEBUG oslo_vmware.api [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Task: {'id': task-794364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143411} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.397208] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.397483] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 752.397769] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.398068] env[61995]: INFO nova.compute.manager [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Took 1.15 seconds to destroy the instance on the hypervisor. [ 752.398342] env[61995]: DEBUG oslo.service.loopingcall [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.398533] env[61995]: DEBUG nova.compute.manager [-] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.398634] env[61995]: DEBUG nova.network.neutron [-] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.475644] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487442} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.476105] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 752.476751] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 752.476751] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cdec0099-5ad8-4434-9c1a-5b521ddba869 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.484317] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 752.484317] env[61995]: value = "task-794368" [ 752.484317] env[61995]: _type = "Task" [ 752.484317] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.495470] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794368, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.695127] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794366, 'name': ReconfigVM_Task, 'duration_secs': 0.3619} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.698280] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Reconfigured VM instance instance-00000025 to attach disk [datastore1] fcdc7c33-8c46-4f4d-bb39-e6f943256cb0/fcdc7c33-8c46-4f4d-bb39-e6f943256cb0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.699483] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f50d4ce4-2d5f-4300-b1df-fb3f5aa61a1f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.705791] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794367, 'name': Rename_Task, 'duration_secs': 0.201642} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.708101] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 752.708432] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 752.708432] env[61995]: value = "task-794369" [ 752.708432] env[61995]: _type = "Task" [ 752.708432] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.709547] env[61995]: DEBUG nova.compute.manager [req-4644d319-69b2-4442-854c-8c2477f119d8 req-b422028d-bdda-43f3-9acb-eced9a3cd0da service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Received event network-vif-deleted-7cc9ac68-fc87-4772-ae4e-041105424b5e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 752.709734] env[61995]: INFO nova.compute.manager [req-4644d319-69b2-4442-854c-8c2477f119d8 req-b422028d-bdda-43f3-9acb-eced9a3cd0da service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Neutron deleted interface 7cc9ac68-fc87-4772-ae4e-041105424b5e; detaching it from the instance and deleting it from the info cache [ 752.709903] env[61995]: DEBUG nova.network.neutron [req-4644d319-69b2-4442-854c-8c2477f119d8 req-b422028d-bdda-43f3-9acb-eced9a3cd0da service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.710945] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26f8ee29-c33a-4995-911c-90d6a837a381 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.724494] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794369, 'name': Rename_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.725071] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 752.725071] env[61995]: value = "task-794370" [ 752.725071] env[61995]: _type = "Task" [ 752.725071] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.733562] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794370, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.842162] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ebde63d-954d-48ee-9a9e-e5f7336a9ce3 tempest-ServerAddressesNegativeTestJSON-548488234 tempest-ServerAddressesNegativeTestJSON-548488234-project-member] Lock "24db5fad-52fd-4689-a8aa-2ae4b0a06710" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.246s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.909611] env[61995]: DEBUG nova.compute.manager [req-ac2f5c03-d452-4b83-92ef-c6db5347dedb req-fff78b2b-da59-49bb-8b7c-5b2eb53bef4a service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Received event network-vif-plugged-a1cabe4c-61ba-460d-b384-421305897add {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 752.909829] env[61995]: DEBUG oslo_concurrency.lockutils [req-ac2f5c03-d452-4b83-92ef-c6db5347dedb req-fff78b2b-da59-49bb-8b7c-5b2eb53bef4a service nova] Acquiring lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.910054] env[61995]: DEBUG oslo_concurrency.lockutils [req-ac2f5c03-d452-4b83-92ef-c6db5347dedb req-fff78b2b-da59-49bb-8b7c-5b2eb53bef4a service nova] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.910235] env[61995]: DEBUG oslo_concurrency.lockutils [req-ac2f5c03-d452-4b83-92ef-c6db5347dedb req-fff78b2b-da59-49bb-8b7c-5b2eb53bef4a service nova] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.910390] env[61995]: DEBUG nova.compute.manager [req-ac2f5c03-d452-4b83-92ef-c6db5347dedb req-fff78b2b-da59-49bb-8b7c-5b2eb53bef4a service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] No waiting events found dispatching network-vif-plugged-a1cabe4c-61ba-460d-b384-421305897add {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 752.910556] env[61995]: WARNING nova.compute.manager [req-ac2f5c03-d452-4b83-92ef-c6db5347dedb req-fff78b2b-da59-49bb-8b7c-5b2eb53bef4a service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Received unexpected event network-vif-plugged-a1cabe4c-61ba-460d-b384-421305897add for instance with vm_state building and task_state spawning. [ 752.985381] env[61995]: DEBUG nova.network.neutron [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Successfully updated port: a1cabe4c-61ba-460d-b384-421305897add {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 752.995693] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794368, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072284} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.996079] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.996919] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6576142b-9b99-47c1-adb8-f459d2e6ce04 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.031180] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.032031] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9146720-71f7-42da-a14b-68a4b1704b81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.063047] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 753.063047] env[61995]: value = "task-794371" [ 753.063047] env[61995]: _type = "Task" [ 753.063047] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.071075] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794371, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.190053] env[61995]: DEBUG nova.network.neutron [-] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.217884] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67fe66a5-4644-44dc-b1e2-3377ec203b97 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.226692] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794369, 'name': Rename_Task, 'duration_secs': 0.246165} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.230373] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 753.230659] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a947d1d-d630-4c78-bc34-0d35ad2e0edf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.234699] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a73a28-8ec9-4c59-82c0-0d268293c626 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.250442] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794370, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.253539] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 753.253539] env[61995]: value = "task-794372" [ 753.253539] env[61995]: _type = "Task" [ 753.253539] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.269595] env[61995]: DEBUG nova.compute.manager [req-4644d319-69b2-4442-854c-8c2477f119d8 req-b422028d-bdda-43f3-9acb-eced9a3cd0da service nova] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Detach interface failed, port_id=7cc9ac68-fc87-4772-ae4e-041105424b5e, reason: Instance a04fbe7e-e53e-43ef-9a6f-ba3895ce7597 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 753.273385] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794372, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.491414] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "refresh_cache-0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.491568] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquired lock "refresh_cache-0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.491865] env[61995]: DEBUG nova.network.neutron [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.572589] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794371, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.692621] env[61995]: INFO nova.compute.manager [-] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Took 1.29 seconds to deallocate network for instance. [ 753.696346] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd33393-7a66-4c79-a234-4d60579b277c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.709307] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cf5d90-c82c-4b24-8270-65c0f0100929 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.747038] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d760d4-b63d-442a-a68d-2be5e2b6b816 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.762514] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794370, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.766767] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb172d6-1501-45a2-adc4-1d50a4f1c979 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.775691] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794372, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.784049] env[61995]: DEBUG nova.compute.provider_tree [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.056808] env[61995]: DEBUG nova.network.neutron [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.073139] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794371, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.209413] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.252974] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794370, 'name': PowerOnVM_Task} progress is 74%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.273212] env[61995]: DEBUG oslo_vmware.api [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794372, 'name': PowerOnVM_Task, 'duration_secs': 0.667636} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.273732] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 754.274614] env[61995]: INFO nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Took 9.49 seconds to spawn the instance on the hypervisor. [ 754.275181] env[61995]: DEBUG nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 754.276105] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2482d768-1566-4cc0-9005-3c10286a8aa7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.286615] env[61995]: DEBUG nova.scheduler.client.report [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.427153] env[61995]: DEBUG nova.network.neutron [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Updating instance_info_cache with network_info: [{"id": "a1cabe4c-61ba-460d-b384-421305897add", "address": "fa:16:3e:11:b4:aa", "network": {"id": "0536e916-9171-4e16-b085-685466cdd990", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-571434529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ec923f4df4f44d19fd832b77dfda303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "67921bdb-a7a0-46b5-ba05-ca997496e222", "external-id": "nsx-vlan-transportzone-856", "segmentation_id": 856, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1cabe4c-61", "ovs_interfaceid": "a1cabe4c-61ba-460d-b384-421305897add", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.573849] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794371, 'name': ReconfigVM_Task, 'duration_secs': 1.026024} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.574172] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Reconfigured VM instance instance-00000027 to attach disk [datastore1] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.574723] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a911aca9-309a-43cf-8759-0976180b2557 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.580898] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 754.580898] env[61995]: value = "task-794373" [ 754.580898] env[61995]: _type = "Task" [ 754.580898] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.588538] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794373, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.754414] env[61995]: DEBUG oslo_vmware.api [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794370, 'name': PowerOnVM_Task, 'duration_secs': 1.848345} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.754704] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 754.754894] env[61995]: INFO nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Took 7.50 seconds to spawn the instance on the hypervisor. [ 754.755079] env[61995]: DEBUG nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 754.755874] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaba72e-7e03-477b-a7ec-a9487d375578 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.799624] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.800136] env[61995]: DEBUG nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.804096] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.768s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.805387] env[61995]: INFO nova.compute.claims [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.809981] env[61995]: INFO nova.compute.manager [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Took 42.10 seconds to build instance. [ 754.933039] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Releasing lock "refresh_cache-0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.933039] env[61995]: DEBUG nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Instance network_info: |[{"id": "a1cabe4c-61ba-460d-b384-421305897add", "address": "fa:16:3e:11:b4:aa", "network": {"id": "0536e916-9171-4e16-b085-685466cdd990", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-571434529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ec923f4df4f44d19fd832b77dfda303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "67921bdb-a7a0-46b5-ba05-ca997496e222", "external-id": "nsx-vlan-transportzone-856", "segmentation_id": 856, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1cabe4c-61", "ovs_interfaceid": "a1cabe4c-61ba-460d-b384-421305897add", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 754.933039] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:b4:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '67921bdb-a7a0-46b5-ba05-ca997496e222', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1cabe4c-61ba-460d-b384-421305897add', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.939800] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Creating folder: Project (9ec923f4df4f44d19fd832b77dfda303). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.939800] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-635da2e5-084a-4cc2-b489-06197d491fe6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.951093] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Created folder: Project (9ec923f4df4f44d19fd832b77dfda303) in parent group-v185203. [ 754.951285] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Creating folder: Instances. Parent ref: group-v185258. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.951517] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68c59b0b-87dd-4413-9989-396dd4921538 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.961849] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Created folder: Instances in parent group-v185258. [ 754.961849] env[61995]: DEBUG oslo.service.loopingcall [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.961849] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 754.962175] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a900f3dc-a2f4-43a2-838f-c89330c7855f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.984651] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.984651] env[61995]: value = "task-794376" [ 754.984651] env[61995]: _type = "Task" [ 754.984651] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.996096] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794376, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.012280] env[61995]: DEBUG nova.compute.manager [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Received event network-changed-a1cabe4c-61ba-460d-b384-421305897add {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 755.012483] env[61995]: DEBUG nova.compute.manager [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Refreshing instance network info cache due to event network-changed-a1cabe4c-61ba-460d-b384-421305897add. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 755.012696] env[61995]: DEBUG oslo_concurrency.lockutils [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] Acquiring lock "refresh_cache-0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.012840] env[61995]: DEBUG oslo_concurrency.lockutils [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] Acquired lock "refresh_cache-0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.013025] env[61995]: DEBUG nova.network.neutron [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Refreshing network info cache for port a1cabe4c-61ba-460d-b384-421305897add {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 755.092055] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794373, 'name': Rename_Task, 'duration_secs': 0.28959} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.092055] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 755.092055] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4fdf371-bb2a-4c06-8c3f-04ac9477d71e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.097246] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 755.097246] env[61995]: value = "task-794377" [ 755.097246] env[61995]: _type = "Task" [ 755.097246] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.104594] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794377, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.274022] env[61995]: INFO nova.compute.manager [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Took 39.31 seconds to build instance. [ 755.312279] env[61995]: DEBUG nova.compute.utils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.313964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d982d87-2296-4b13-80f6-3df310b1dbbe tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.638s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.314238] env[61995]: DEBUG nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.314399] env[61995]: DEBUG nova.network.neutron [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.393309] env[61995]: DEBUG nova.policy [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c135f484cfe0401f83d1fcf16debf8b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a7fbb0e40964f349256c7529874ca8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 755.495108] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794376, 'name': CreateVM_Task, 'duration_secs': 0.504186} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.495305] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 755.496080] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.496342] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.496668] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.496930] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab7d1206-63fe-4e83-ad7d-fc3d6386e47e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.503688] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 755.503688] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527c1eb5-222c-e10d-0bcd-6641d528bbea" [ 755.503688] env[61995]: _type = "Task" [ 755.503688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.512031] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527c1eb5-222c-e10d-0bcd-6641d528bbea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.610021] env[61995]: DEBUG oslo_vmware.api [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794377, 'name': PowerOnVM_Task, 'duration_secs': 0.502634} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.610021] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 755.610236] env[61995]: INFO nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Took 5.83 seconds to spawn the instance on the hypervisor. [ 755.610487] env[61995]: DEBUG nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.611270] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de6e4a3-b73a-4169-bb0d-9fa8a04da45a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.749161] env[61995]: DEBUG nova.network.neutron [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Successfully created port: 591f35b8-baf5-4da1-b479-dc99ae2f7b75 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.775498] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8da0e172-36b1-47fa-b015-b62fcf9ec7d2 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "55555deb-9883-483b-a901-42b1a66c60d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.328s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.785767] env[61995]: DEBUG nova.network.neutron [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Updated VIF entry in instance network info cache for port a1cabe4c-61ba-460d-b384-421305897add. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 755.786178] env[61995]: DEBUG nova.network.neutron [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Updating instance_info_cache with network_info: [{"id": "a1cabe4c-61ba-460d-b384-421305897add", "address": "fa:16:3e:11:b4:aa", "network": {"id": "0536e916-9171-4e16-b085-685466cdd990", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-571434529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ec923f4df4f44d19fd832b77dfda303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "67921bdb-a7a0-46b5-ba05-ca997496e222", "external-id": "nsx-vlan-transportzone-856", "segmentation_id": 856, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1cabe4c-61", "ovs_interfaceid": "a1cabe4c-61ba-460d-b384-421305897add", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.818732] env[61995]: DEBUG nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.822105] env[61995]: DEBUG nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.967205] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.967620] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.015684] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527c1eb5-222c-e10d-0bcd-6641d528bbea, 'name': SearchDatastore_Task, 'duration_secs': 0.011518} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.018510] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.018729] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 756.018916] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.019139] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.019338] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.019790] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d4c93e4-b174-41f2-8743-cc1647dd4cb9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.029452] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 756.029618] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 756.030537] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65ed86cc-e559-4013-9a40-547626f124c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.040350] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 756.040350] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52848460-4a13-904f-6e3d-ae66fc83589d" [ 756.040350] env[61995]: _type = "Task" [ 756.040350] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.049410] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52848460-4a13-904f-6e3d-ae66fc83589d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.139451] env[61995]: INFO nova.compute.manager [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Took 34.09 seconds to build instance. [ 756.259382] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077431a3-a19f-4fa3-96c4-6aecf206ef07 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.268792] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d35e23-7dd7-4fc4-991c-3ef5ad2623f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.299721] env[61995]: DEBUG nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 756.306212] env[61995]: DEBUG oslo_concurrency.lockutils [req-4b1f054c-4e8b-4a12-b6fb-d0148368d2a4 req-3b6fc37b-8182-4b0a-a70c-4c4c69c2f23c service nova] Releasing lock "refresh_cache-0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.308399] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557dc4ee-a26e-4911-a0ad-2bee3eefd3dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.321356] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda698dc-6929-4d7c-bb61-44dab0903ab3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.348530] env[61995]: DEBUG nova.compute.provider_tree [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.361324] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.552585] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52848460-4a13-904f-6e3d-ae66fc83589d, 'name': SearchDatastore_Task, 'duration_secs': 0.008851} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.553613] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51964129-291b-4b02-b31a-51138d49e546 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.559471] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 756.559471] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5242342b-1e11-a677-61d1-0b368b9c7132" [ 756.559471] env[61995]: _type = "Task" [ 756.559471] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.567866] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5242342b-1e11-a677-61d1-0b368b9c7132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.641957] env[61995]: DEBUG oslo_concurrency.lockutils [None req-26fc53c4-cdff-4a7a-b906-77376ef5272d tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.420s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.815470] env[61995]: INFO nova.compute.manager [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Rebuilding instance [ 756.818351] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.834321] env[61995]: DEBUG nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.851864] env[61995]: DEBUG nova.scheduler.client.report [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.860044] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.860044] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.860044] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.860044] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.860044] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.860379] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.860379] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.860479] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.860641] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.860832] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.860967] env[61995]: DEBUG nova.virt.hardware [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.861962] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e24e66d-3a15-430b-a1e3-252d164bb3ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.865381] env[61995]: DEBUG nova.compute.manager [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 756.866192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2ef773-cece-4aa7-b818-52cf197e877e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.875892] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821d1914-52f0-48e0-b85e-6adfc389b443 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.070108] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5242342b-1e11-a677-61d1-0b368b9c7132, 'name': SearchDatastore_Task, 'duration_secs': 0.009537} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.070367] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.070625] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa/0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 757.070872] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0300644-17fc-49fd-952f-5c97dfc919c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.078079] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 757.078079] env[61995]: value = "task-794378" [ 757.078079] env[61995]: _type = "Task" [ 757.078079] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.086429] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794378, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.144420] env[61995]: DEBUG nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.291751] env[61995]: DEBUG nova.compute.manager [req-8e43cff2-db46-4e76-9438-119262e76470 req-82323015-98e8-4b5e-ab99-4afb7f821c5f service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Received event network-vif-plugged-591f35b8-baf5-4da1-b479-dc99ae2f7b75 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 757.291991] env[61995]: DEBUG oslo_concurrency.lockutils [req-8e43cff2-db46-4e76-9438-119262e76470 req-82323015-98e8-4b5e-ab99-4afb7f821c5f service nova] Acquiring lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.292241] env[61995]: DEBUG oslo_concurrency.lockutils [req-8e43cff2-db46-4e76-9438-119262e76470 req-82323015-98e8-4b5e-ab99-4afb7f821c5f service nova] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.292452] env[61995]: DEBUG oslo_concurrency.lockutils [req-8e43cff2-db46-4e76-9438-119262e76470 req-82323015-98e8-4b5e-ab99-4afb7f821c5f service nova] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.292636] env[61995]: DEBUG nova.compute.manager [req-8e43cff2-db46-4e76-9438-119262e76470 req-82323015-98e8-4b5e-ab99-4afb7f821c5f service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] No waiting events found dispatching network-vif-plugged-591f35b8-baf5-4da1-b479-dc99ae2f7b75 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 757.292934] env[61995]: WARNING nova.compute.manager [req-8e43cff2-db46-4e76-9438-119262e76470 req-82323015-98e8-4b5e-ab99-4afb7f821c5f service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Received unexpected event network-vif-plugged-591f35b8-baf5-4da1-b479-dc99ae2f7b75 for instance with vm_state building and task_state spawning. [ 757.366387] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.367015] env[61995]: DEBUG nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.369999] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.054s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.371694] env[61995]: INFO nova.compute.claims [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.383721] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 757.384043] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd84e337-a1d4-4baa-a305-42d20c6935f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.394399] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 757.394399] env[61995]: value = "task-794379" [ 757.394399] env[61995]: _type = "Task" [ 757.394399] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.405847] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794379, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.453933] env[61995]: DEBUG nova.network.neutron [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Successfully updated port: 591f35b8-baf5-4da1-b479-dc99ae2f7b75 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 757.591314] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794378, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451718} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.591314] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa/0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 757.591314] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.591314] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e655be9-dfb6-46e4-bfc7-4e4c911e7076 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.596446] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 757.596446] env[61995]: value = "task-794380" [ 757.596446] env[61995]: _type = "Task" [ 757.596446] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.607247] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.666785] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.876632] env[61995]: DEBUG nova.compute.utils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.880674] env[61995]: DEBUG nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.880848] env[61995]: DEBUG nova.network.neutron [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 757.905271] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794379, 'name': PowerOffVM_Task, 'duration_secs': 0.165178} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.905542] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 757.905759] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 757.906533] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289887af-c643-419b-bc8c-95b4415e96a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.913494] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 757.913718] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43ea9878-ed27-456a-b5cb-f02139ded4c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.926624] env[61995]: DEBUG nova.policy [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07810f995add491783ea409b404ef51d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5a84fcd26824a0bbd2a65440d615a1c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 757.941403] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 757.941607] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 757.941787] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleting the datastore file [datastore1] ecb72668-caa4-4117-9f8d-e0c6e4831f3d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 757.942045] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a0a3cac-cc4e-4c00-8f1c-39d81a947935 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.948871] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 757.948871] env[61995]: value = "task-794382" [ 757.948871] env[61995]: _type = "Task" [ 757.948871] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.957543] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.957683] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.957822] env[61995]: DEBUG nova.network.neutron [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.959558] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794382, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.107513] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094983} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.107802] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.108700] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab88657-1ebd-4d16-9651-417426bb7ed2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.131470] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa/0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.132515] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2db9f89f-2164-4da2-be1f-cdaba3dc3e01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.151567] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 758.151567] env[61995]: value = "task-794383" [ 758.151567] env[61995]: _type = "Task" [ 758.151567] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.163026] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794383, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.246084] env[61995]: DEBUG nova.network.neutron [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Successfully created port: b040835e-914d-428e-9341-ee8c08259624 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.386090] env[61995]: DEBUG nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.468080] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794382, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109865} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.468661] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 758.468903] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 758.470217] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 758.518607] env[61995]: DEBUG nova.network.neutron [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.663090] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794383, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.694333] env[61995]: DEBUG nova.network.neutron [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance_info_cache with network_info: [{"id": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "address": "fa:16:3e:3d:94:75", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap591f35b8-ba", "ovs_interfaceid": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.765716] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a1b311-8bed-48d0-a00c-c2f78e2b7180 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.773396] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f41e85-4db4-4a70-911f-34c38f89d75c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.803883] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2b3b62-8f86-42c3-906a-848b1b44c726 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.810099] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b8d9d5-94e0-46b7-8163-40a1a03419f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.823115] env[61995]: DEBUG nova.compute.provider_tree [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.161770] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794383, 'name': ReconfigVM_Task, 'duration_secs': 0.838052} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.162076] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa/0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 759.162711] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad962347-fd7d-46d7-9611-2e237be3df17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.169495] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 759.169495] env[61995]: value = "task-794384" [ 759.169495] env[61995]: _type = "Task" [ 759.169495] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.178139] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794384, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.200127] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.200490] env[61995]: DEBUG nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Instance network_info: |[{"id": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "address": "fa:16:3e:3d:94:75", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap591f35b8-ba", "ovs_interfaceid": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 759.200903] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:94:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '591f35b8-baf5-4da1-b479-dc99ae2f7b75', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.208335] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Creating folder: Project (5a7fbb0e40964f349256c7529874ca8e). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.208606] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f601d5b-eb10-4968-a0c0-2fbbbde7a975 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.218619] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Created folder: Project (5a7fbb0e40964f349256c7529874ca8e) in parent group-v185203. [ 759.218813] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Creating folder: Instances. Parent ref: group-v185261. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.219059] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dfe46ead-92e4-48d6-8f0c-028b1526c7fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.228347] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Created folder: Instances in parent group-v185261. [ 759.228623] env[61995]: DEBUG oslo.service.loopingcall [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.228785] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 759.228948] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0365d0db-826c-4ed5-bd10-e0145dba92ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.248794] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 759.248794] env[61995]: value = "task-794387" [ 759.248794] env[61995]: _type = "Task" [ 759.248794] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.255670] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794387, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.324196] env[61995]: DEBUG nova.compute.manager [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Received event network-changed-591f35b8-baf5-4da1-b479-dc99ae2f7b75 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 759.324329] env[61995]: DEBUG nova.compute.manager [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Refreshing instance network info cache due to event network-changed-591f35b8-baf5-4da1-b479-dc99ae2f7b75. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 759.324544] env[61995]: DEBUG oslo_concurrency.lockutils [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] Acquiring lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.324684] env[61995]: DEBUG oslo_concurrency.lockutils [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] Acquired lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.324840] env[61995]: DEBUG nova.network.neutron [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Refreshing network info cache for port 591f35b8-baf5-4da1-b479-dc99ae2f7b75 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 759.326876] env[61995]: DEBUG nova.scheduler.client.report [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.395983] env[61995]: DEBUG nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.421481] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.421755] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.421943] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.422176] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.422365] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.422536] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.422773] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.422968] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.423192] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.423390] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.423593] env[61995]: DEBUG nova.virt.hardware [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.424516] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8c40e9-641a-458e-920f-799f53e809f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.432567] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b727dbdd-68a9-4857-84ef-b301c3d24291 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.502983] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.502983] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.502983] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.503230] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.503273] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.503401] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.503607] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.503765] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.503929] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.504103] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.504279] env[61995]: DEBUG nova.virt.hardware [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.505134] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282f7f9f-10cd-43ef-8dbc-deefe04c3235 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.513537] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef84debe-3972-4ca7-8c4f-34a650a31b8d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.527988] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.533157] env[61995]: DEBUG oslo.service.loopingcall [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.533423] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 759.533630] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e19c3b3d-e2fa-4164-a285-d66cddc23121 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.549487] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 759.549487] env[61995]: value = "task-794388" [ 759.549487] env[61995]: _type = "Task" [ 759.549487] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.558885] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794388, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.679358] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794384, 'name': Rename_Task, 'duration_secs': 0.150749} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.679632] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 759.679862] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-505681f4-c95a-4079-87bc-825b1c9bddbe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.685590] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 759.685590] env[61995]: value = "task-794389" [ 759.685590] env[61995]: _type = "Task" [ 759.685590] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.696170] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.757753] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794387, 'name': CreateVM_Task, 'duration_secs': 0.321065} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.757962] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 759.758696] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.758858] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.759232] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 759.759481] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87809096-be53-44fd-8afe-dff492bb8bc4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.764219] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 759.764219] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5298eadd-d975-0695-7cbc-32224275e8c4" [ 759.764219] env[61995]: _type = "Task" [ 759.764219] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.771963] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5298eadd-d975-0695-7cbc-32224275e8c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.821076] env[61995]: DEBUG nova.network.neutron [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Successfully updated port: b040835e-914d-428e-9341-ee8c08259624 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 759.834142] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.834664] env[61995]: DEBUG nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.838563] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.753s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.838784] env[61995]: DEBUG nova.objects.instance [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lazy-loading 'resources' on Instance uuid 1c4eb398-c4e9-4588-ba48-2805e3fee750 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 760.059815] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794388, 'name': CreateVM_Task, 'duration_secs': 0.256334} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.060044] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 760.060515] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.144905] env[61995]: DEBUG nova.network.neutron [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updated VIF entry in instance network info cache for port 591f35b8-baf5-4da1-b479-dc99ae2f7b75. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 760.144905] env[61995]: DEBUG nova.network.neutron [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance_info_cache with network_info: [{"id": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "address": "fa:16:3e:3d:94:75", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap591f35b8-ba", "ovs_interfaceid": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.197792] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794389, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.273841] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5298eadd-d975-0695-7cbc-32224275e8c4, 'name': SearchDatastore_Task, 'duration_secs': 0.008687} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.274167] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.274408] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.274640] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.274782] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.274958] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 760.275880] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.275880] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 760.275880] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed461742-af66-4718-8815-2662defb15b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.277626] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdd355cf-7c9e-4c4c-8733-f9b80a2b0992 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.282457] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 760.282457] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5276a623-059d-c6b8-44a1-46d9fa7cc196" [ 760.282457] env[61995]: _type = "Task" [ 760.282457] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.286699] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 760.286878] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 760.288169] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd92d547-13e1-4937-bc67-999a1df28571 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.292797] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5276a623-059d-c6b8-44a1-46d9fa7cc196, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.295593] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 760.295593] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e3a529-ee22-444f-3378-23e313baf042" [ 760.295593] env[61995]: _type = "Task" [ 760.295593] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.303021] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e3a529-ee22-444f-3378-23e313baf042, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.326825] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "refresh_cache-4da426a1-7af9-4e4e-82ed-c45687a19b86" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.326958] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquired lock "refresh_cache-4da426a1-7af9-4e4e-82ed-c45687a19b86" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.327133] env[61995]: DEBUG nova.network.neutron [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.339994] env[61995]: DEBUG nova.compute.utils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.342437] env[61995]: DEBUG nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.342613] env[61995]: DEBUG nova.network.neutron [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 760.386075] env[61995]: DEBUG nova.policy [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93b1d26f7a03444d8c7fdab6fac57ea2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '556cd4687d78408bbd5e53e3aec7eecc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 760.634398] env[61995]: DEBUG nova.network.neutron [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Successfully created port: fac419a3-c8d4-48b4-8dde-b6a64593a275 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.647603] env[61995]: DEBUG oslo_concurrency.lockutils [req-679678e8-a320-4fac-af82-0e05b020bedd req-72b3335c-135b-4d05-be94-d702ab37a50b service nova] Releasing lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.657719] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aadb072-9879-4f39-b404-00ed686e5d79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.668019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd782443-59df-4715-bf7e-bfedb86d851b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.698295] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56f2c3e-c9b8-4ede-bd6f-bb7e9ae7dae1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.705950] env[61995]: DEBUG oslo_vmware.api [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794389, 'name': PowerOnVM_Task, 'duration_secs': 0.558881} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.708165] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 760.708345] env[61995]: INFO nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Took 8.46 seconds to spawn the instance on the hypervisor. [ 760.708934] env[61995]: DEBUG nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.709341] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593d253d-033f-4572-ade2-f94f02f52c61 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.712617] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5adbb73-24af-4522-94fe-a44d2acc9b28 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.733445] env[61995]: DEBUG nova.compute.provider_tree [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.792750] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5276a623-059d-c6b8-44a1-46d9fa7cc196, 'name': SearchDatastore_Task, 'duration_secs': 0.009413} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.793093] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.793298] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.793501] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.804921] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e3a529-ee22-444f-3378-23e313baf042, 'name': SearchDatastore_Task, 'duration_secs': 0.008096} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.805656] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-169954ff-0cfc-4b47-ac0f-6d9fba28434c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.810591] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 760.810591] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521a483d-a7b9-af81-84e5-86cd386490cd" [ 760.810591] env[61995]: _type = "Task" [ 760.810591] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.817896] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521a483d-a7b9-af81-84e5-86cd386490cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.843292] env[61995]: DEBUG nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.883387] env[61995]: DEBUG nova.network.neutron [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.047591] env[61995]: DEBUG nova.network.neutron [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Updating instance_info_cache with network_info: [{"id": "b040835e-914d-428e-9341-ee8c08259624", "address": "fa:16:3e:24:97:01", "network": {"id": "0d37904f-3a5e-411f-ae3c-f406bea2908e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-480645540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5a84fcd26824a0bbd2a65440d615a1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb040835e-91", "ovs_interfaceid": "b040835e-914d-428e-9341-ee8c08259624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.240490] env[61995]: DEBUG nova.scheduler.client.report [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.248045] env[61995]: INFO nova.compute.manager [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Took 38.14 seconds to build instance. [ 761.321476] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521a483d-a7b9-af81-84e5-86cd386490cd, 'name': SearchDatastore_Task, 'duration_secs': 0.009148} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.321736] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.322134] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 4edbf65d-eaca-47e1-b6bf-03ccb908b52e/4edbf65d-eaca-47e1-b6bf-03ccb908b52e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 761.322444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.322632] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 761.322840] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-172e3d68-c9ba-418c-9607-7ce461dbcda2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.325150] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0eb88384-46e0-440a-bc56-c59e3ca268a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.331957] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 761.331957] env[61995]: value = "task-794390" [ 761.331957] env[61995]: _type = "Task" [ 761.331957] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.333256] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 761.333553] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 761.336890] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-560f9a8d-b3c3-4930-bc46-1c567a938ef6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.344399] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794390, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.346037] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 761.346037] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dfc81d-1ab9-2594-4304-0914c2fc0d9a" [ 761.346037] env[61995]: _type = "Task" [ 761.346037] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.357853] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52dfc81d-1ab9-2594-4304-0914c2fc0d9a, 'name': SearchDatastore_Task, 'duration_secs': 0.008971} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.361507] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4efdadca-5357-4eb1-bbe9-fe9f8636fc15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.368288] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 761.368288] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5254bc14-8db1-3048-1319-2ebf24a9ec33" [ 761.368288] env[61995]: _type = "Task" [ 761.368288] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.377265] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5254bc14-8db1-3048-1319-2ebf24a9ec33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.429983] env[61995]: DEBUG nova.compute.manager [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Received event network-vif-plugged-b040835e-914d-428e-9341-ee8c08259624 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 761.430279] env[61995]: DEBUG oslo_concurrency.lockutils [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] Acquiring lock "4da426a1-7af9-4e4e-82ed-c45687a19b86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.430880] env[61995]: DEBUG oslo_concurrency.lockutils [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.430880] env[61995]: DEBUG oslo_concurrency.lockutils [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.431069] env[61995]: DEBUG nova.compute.manager [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] No waiting events found dispatching network-vif-plugged-b040835e-914d-428e-9341-ee8c08259624 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 761.431206] env[61995]: WARNING nova.compute.manager [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Received unexpected event network-vif-plugged-b040835e-914d-428e-9341-ee8c08259624 for instance with vm_state building and task_state spawning. [ 761.431451] env[61995]: DEBUG nova.compute.manager [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Received event network-changed-b040835e-914d-428e-9341-ee8c08259624 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 761.431662] env[61995]: DEBUG nova.compute.manager [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Refreshing instance network info cache due to event network-changed-b040835e-914d-428e-9341-ee8c08259624. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 761.431884] env[61995]: DEBUG oslo_concurrency.lockutils [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] Acquiring lock "refresh_cache-4da426a1-7af9-4e4e-82ed-c45687a19b86" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.550545] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Releasing lock "refresh_cache-4da426a1-7af9-4e4e-82ed-c45687a19b86" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.550993] env[61995]: DEBUG nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Instance network_info: |[{"id": "b040835e-914d-428e-9341-ee8c08259624", "address": "fa:16:3e:24:97:01", "network": {"id": "0d37904f-3a5e-411f-ae3c-f406bea2908e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-480645540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5a84fcd26824a0bbd2a65440d615a1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb040835e-91", "ovs_interfaceid": "b040835e-914d-428e-9341-ee8c08259624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 761.551391] env[61995]: DEBUG oslo_concurrency.lockutils [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] Acquired lock "refresh_cache-4da426a1-7af9-4e4e-82ed-c45687a19b86" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.551600] env[61995]: DEBUG nova.network.neutron [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Refreshing network info cache for port b040835e-914d-428e-9341-ee8c08259624 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 761.553110] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:97:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b040835e-914d-428e-9341-ee8c08259624', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.561474] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Creating folder: Project (b5a84fcd26824a0bbd2a65440d615a1c). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 761.563514] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ad532f4-f26c-4801-b188-114a485daff1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.576284] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Created folder: Project (b5a84fcd26824a0bbd2a65440d615a1c) in parent group-v185203. [ 761.576546] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Creating folder: Instances. Parent ref: group-v185265. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 761.576775] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a0139ba-f8c4-497a-b25c-dd257b4cfd96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.590317] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Created folder: Instances in parent group-v185265. [ 761.590620] env[61995]: DEBUG oslo.service.loopingcall [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.590830] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 761.591098] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-85451929-58b6-4ad1-94d2-31e9cd1413bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.613086] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.613086] env[61995]: value = "task-794393" [ 761.613086] env[61995]: _type = "Task" [ 761.613086] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.622511] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794393, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.749102] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.910s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.751924] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.488s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.752161] env[61995]: DEBUG nova.objects.instance [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 761.755411] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8869c217-6b10-454f-b004-71973fd6c8bd tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.399s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.780067] env[61995]: INFO nova.scheduler.client.report [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Deleted allocations for instance 1c4eb398-c4e9-4588-ba48-2805e3fee750 [ 761.845843] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794390, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444058} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.846120] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 4edbf65d-eaca-47e1-b6bf-03ccb908b52e/4edbf65d-eaca-47e1-b6bf-03ccb908b52e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 761.846365] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 761.846579] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb8c3dd1-d4b3-4193-afeb-a7c6c9cbf6bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.853919] env[61995]: DEBUG nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.857017] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 761.857017] env[61995]: value = "task-794394" [ 761.857017] env[61995]: _type = "Task" [ 761.857017] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.865344] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.865715] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.865990] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.866252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.866501] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.868763] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.869813] env[61995]: INFO nova.compute.manager [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Terminating instance [ 761.875081] env[61995]: DEBUG nova.compute.manager [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.875280] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 761.877748] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e1e6e4-f20c-40f5-bad7-a19b8fa58269 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.886097] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5254bc14-8db1-3048-1319-2ebf24a9ec33, 'name': SearchDatastore_Task, 'duration_secs': 0.008953} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.888339] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.888471] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.888603] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.888782] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.888925] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.889084] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.889305] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.889451] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.889610] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.889766] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.889938] env[61995]: DEBUG nova.virt.hardware [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.892117] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.892371] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 761.893306] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5890bbc7-dd2f-4cdb-b834-147e94a5b793 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.895911] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 761.896358] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db350bdc-ff59-4e57-ac2b-1a00d5500429 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.898370] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8281d22-1eef-416f-8108-de29242ddc59 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.907262] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0545512f-8e12-4cce-b32d-a2de2af7210a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.911680] env[61995]: DEBUG oslo_vmware.api [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 761.911680] env[61995]: value = "task-794396" [ 761.911680] env[61995]: _type = "Task" [ 761.911680] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.911975] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 761.911975] env[61995]: value = "task-794395" [ 761.911975] env[61995]: _type = "Task" [ 761.911975] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.934862] env[61995]: DEBUG oslo_vmware.api [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.935134] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794395, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.125557] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794393, 'name': CreateVM_Task, 'duration_secs': 0.42426} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.127874] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 762.128843] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.129082] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.129454] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.130152] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5654ac3-b6b4-4406-8006-014f9913411b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.136256] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 762.136256] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52895af7-780e-1279-99e5-8e629d5f9b2a" [ 762.136256] env[61995]: _type = "Task" [ 762.136256] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.145801] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52895af7-780e-1279-99e5-8e629d5f9b2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.262713] env[61995]: DEBUG nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.289485] env[61995]: DEBUG oslo_concurrency.lockutils [None req-00a88d0f-4d8d-46fe-83a4-964c7a27a810 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "1c4eb398-c4e9-4588-ba48-2805e3fee750" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.868s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.367138] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073343} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.367454] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.368726] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b34b60-8161-411e-bcdf-ad29a5a0d701 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.396703] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 4edbf65d-eaca-47e1-b6bf-03ccb908b52e/4edbf65d-eaca-47e1-b6bf-03ccb908b52e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.397103] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37bbbfc0-a6af-4d01-ba7f-2b574160a178 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.414093] env[61995]: DEBUG nova.network.neutron [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Updated VIF entry in instance network info cache for port b040835e-914d-428e-9341-ee8c08259624. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 762.414484] env[61995]: DEBUG nova.network.neutron [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Updating instance_info_cache with network_info: [{"id": "b040835e-914d-428e-9341-ee8c08259624", "address": "fa:16:3e:24:97:01", "network": {"id": "0d37904f-3a5e-411f-ae3c-f406bea2908e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-480645540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5a84fcd26824a0bbd2a65440d615a1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb040835e-91", "ovs_interfaceid": "b040835e-914d-428e-9341-ee8c08259624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.430116] env[61995]: DEBUG oslo_vmware.api [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794396, 'name': PowerOffVM_Task, 'duration_secs': 0.180041} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.434534] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 762.436020] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 762.436020] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794395, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488548} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.436020] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 762.436020] env[61995]: value = "task-794397" [ 762.436020] env[61995]: _type = "Task" [ 762.436020] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.436289] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2da0ea5-f902-47cf-80df-f1c3baef849b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.437887] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 762.438230] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.438781] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dad83f28-914b-42a5-8a48-161023c5e8ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.450683] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794397, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.452207] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 762.452207] env[61995]: value = "task-794399" [ 762.452207] env[61995]: _type = "Task" [ 762.452207] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.462883] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794399, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.532851] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 762.532851] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 762.532851] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Deleting the datastore file [datastore2] 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.532851] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a98a154-398c-4845-8886-90cabbd57f6b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.538976] env[61995]: DEBUG oslo_vmware.api [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for the task: (returnval){ [ 762.538976] env[61995]: value = "task-794400" [ 762.538976] env[61995]: _type = "Task" [ 762.538976] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.548511] env[61995]: DEBUG oslo_vmware.api [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.652818] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52895af7-780e-1279-99e5-8e629d5f9b2a, 'name': SearchDatastore_Task, 'duration_secs': 0.061144} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.652818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.652818] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.652818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.653163] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.653163] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.653402] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fe4d965-e547-45a8-ad17-289fc423e3b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.662886] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.663075] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 762.663774] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bdedd8e-9968-43c0-a068-fb2a81f52432 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.669328] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 762.669328] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a2f5e1-8f9c-b6e3-04a6-4e96dc3b8a6e" [ 762.669328] env[61995]: _type = "Task" [ 762.669328] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.677034] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a2f5e1-8f9c-b6e3-04a6-4e96dc3b8a6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.747909] env[61995]: DEBUG nova.network.neutron [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Successfully updated port: fac419a3-c8d4-48b4-8dde-b6a64593a275 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 762.771454] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6c16f1e5-f977-436f-b742-9a1520479d81 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.775537] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.291s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.776964] env[61995]: INFO nova.compute.claims [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.798241] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.926027] env[61995]: DEBUG oslo_concurrency.lockutils [req-d9fdfaee-54b2-4c22-90d2-5f3d2e84bd25 req-b3b5d271-52fe-42f8-b55b-478023b5cc06 service nova] Releasing lock "refresh_cache-4da426a1-7af9-4e4e-82ed-c45687a19b86" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.949698] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794397, 'name': ReconfigVM_Task, 'duration_secs': 0.300851} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.949990] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 4edbf65d-eaca-47e1-b6bf-03ccb908b52e/4edbf65d-eaca-47e1-b6bf-03ccb908b52e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 762.950839] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f984579-66ae-4690-ae61-f7ab02cd5913 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.964251] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794399, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114831} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.965567] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.965908] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 762.965908] env[61995]: value = "task-794401" [ 762.965908] env[61995]: _type = "Task" [ 762.965908] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.966622] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce54d78d-4e4b-419c-b333-c16a4d1393fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.977790] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794401, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.997392] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.997778] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6a378e0-cd86-4db3-bd0d-6dfaa3a6240f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.019420] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 763.019420] env[61995]: value = "task-794402" [ 763.019420] env[61995]: _type = "Task" [ 763.019420] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.028127] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794402, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.047690] env[61995]: DEBUG oslo_vmware.api [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Task: {'id': task-794400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.413478} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.047888] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.048868] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 763.048868] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 763.048868] env[61995]: INFO nova.compute.manager [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Took 1.17 seconds to destroy the instance on the hypervisor. [ 763.048868] env[61995]: DEBUG oslo.service.loopingcall [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.048868] env[61995]: DEBUG nova.compute.manager [-] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.049168] env[61995]: DEBUG nova.network.neutron [-] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.179234] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a2f5e1-8f9c-b6e3-04a6-4e96dc3b8a6e, 'name': SearchDatastore_Task, 'duration_secs': 0.024931} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.183013] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7799019-2b92-4e7a-a811-592e0ff9d160 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.185055] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 763.185055] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5276d79e-5530-8534-e6f9-78e4d1bd94cf" [ 763.185055] env[61995]: _type = "Task" [ 763.185055] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.192583] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5276d79e-5530-8534-e6f9-78e4d1bd94cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.251097] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "refresh_cache-2da17a43-2260-4a82-9ed3-eafd25bb9a6a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.251097] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquired lock "refresh_cache-2da17a43-2260-4a82-9ed3-eafd25bb9a6a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.251097] env[61995]: DEBUG nova.network.neutron [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.480273] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794401, 'name': Rename_Task, 'duration_secs': 0.338459} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.481015] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 763.481447] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-126a8b7f-61bb-45b6-9847-b5a6d5afd5ce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.490698] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 763.490698] env[61995]: value = "task-794403" [ 763.490698] env[61995]: _type = "Task" [ 763.490698] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.498488] env[61995]: DEBUG nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Received event network-vif-plugged-fac419a3-c8d4-48b4-8dde-b6a64593a275 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.499261] env[61995]: DEBUG oslo_concurrency.lockutils [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] Acquiring lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.499386] env[61995]: DEBUG oslo_concurrency.lockutils [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.499663] env[61995]: DEBUG oslo_concurrency.lockutils [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.500028] env[61995]: DEBUG nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] No waiting events found dispatching network-vif-plugged-fac419a3-c8d4-48b4-8dde-b6a64593a275 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 763.500628] env[61995]: WARNING nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Received unexpected event network-vif-plugged-fac419a3-c8d4-48b4-8dde-b6a64593a275 for instance with vm_state building and task_state spawning. [ 763.501335] env[61995]: DEBUG nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Received event network-changed-fac419a3-c8d4-48b4-8dde-b6a64593a275 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.501335] env[61995]: DEBUG nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Refreshing instance network info cache due to event network-changed-fac419a3-c8d4-48b4-8dde-b6a64593a275. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 763.501827] env[61995]: DEBUG oslo_concurrency.lockutils [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] Acquiring lock "refresh_cache-2da17a43-2260-4a82-9ed3-eafd25bb9a6a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.507586] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794403, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.532764] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794402, 'name': ReconfigVM_Task, 'duration_secs': 0.319079} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.534707] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Reconfigured VM instance instance-00000027 to attach disk [datastore2] ecb72668-caa4-4117-9f8d-e0c6e4831f3d/ecb72668-caa4-4117-9f8d-e0c6e4831f3d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.535341] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f72984a-3114-4ccd-ac92-d208ffc311bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.543053] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 763.543053] env[61995]: value = "task-794404" [ 763.543053] env[61995]: _type = "Task" [ 763.543053] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.558302] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794404, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.694754] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5276d79e-5530-8534-e6f9-78e4d1bd94cf, 'name': SearchDatastore_Task, 'duration_secs': 0.01001} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.695087] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.695284] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 4da426a1-7af9-4e4e-82ed-c45687a19b86/4da426a1-7af9-4e4e-82ed-c45687a19b86.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 763.695539] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ab1cce0-a52a-4359-879e-cfdee2d56d5c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.701526] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 763.701526] env[61995]: value = "task-794405" [ 763.701526] env[61995]: _type = "Task" [ 763.701526] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.709826] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794405, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.808801] env[61995]: DEBUG nova.network.neutron [-] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.006753] env[61995]: DEBUG nova.network.neutron [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.018887] env[61995]: DEBUG oslo_vmware.api [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794403, 'name': PowerOnVM_Task, 'duration_secs': 0.513354} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.019248] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 764.019483] env[61995]: INFO nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Took 7.18 seconds to spawn the instance on the hypervisor. [ 764.019625] env[61995]: DEBUG nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 764.020506] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4bd27a-4edb-4670-a400-8f21099637e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.057594] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794404, 'name': Rename_Task, 'duration_secs': 0.163434} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.057982] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 764.058293] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6f2dbb7-77fe-4d2d-b682-3f115e514c80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.070960] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 764.070960] env[61995]: value = "task-794406" [ 764.070960] env[61995]: _type = "Task" [ 764.070960] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.079903] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.211325] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794405, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.241370] env[61995]: DEBUG nova.network.neutron [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Updating instance_info_cache with network_info: [{"id": "fac419a3-c8d4-48b4-8dde-b6a64593a275", "address": "fa:16:3e:dc:68:18", "network": {"id": "b9bf6713-8a48-4ec9-ab6f-4972939c5daf", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-889552615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "556cd4687d78408bbd5e53e3aec7eecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfac419a3-c8", "ovs_interfaceid": "fac419a3-c8d4-48b4-8dde-b6a64593a275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.251788] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcb17d0-8f28-45d8-9345-a0beb27cd633 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.259805] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f0fd54-9c73-4706-a5ed-0f2bd5d5b5f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.293807] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06db35d0-9568-42d1-ac3d-9505a976703f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.301241] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd9d09a-be5e-4cad-b3d7-37ddc73938e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.319240] env[61995]: INFO nova.compute.manager [-] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Took 1.27 seconds to deallocate network for instance. [ 764.320147] env[61995]: DEBUG nova.compute.provider_tree [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.546582] env[61995]: INFO nova.compute.manager [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Took 38.97 seconds to build instance. [ 764.583765] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794406, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.711830] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794405, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51127} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.712121] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 4da426a1-7af9-4e4e-82ed-c45687a19b86/4da426a1-7af9-4e4e-82ed-c45687a19b86.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 764.712344] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 764.712593] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd0aaa08-97e4-442d-be21-d9ecd1623d5d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.718878] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 764.718878] env[61995]: value = "task-794407" [ 764.718878] env[61995]: _type = "Task" [ 764.718878] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.726806] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.743556] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Releasing lock "refresh_cache-2da17a43-2260-4a82-9ed3-eafd25bb9a6a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.743886] env[61995]: DEBUG nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Instance network_info: |[{"id": "fac419a3-c8d4-48b4-8dde-b6a64593a275", "address": "fa:16:3e:dc:68:18", "network": {"id": "b9bf6713-8a48-4ec9-ab6f-4972939c5daf", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-889552615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "556cd4687d78408bbd5e53e3aec7eecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfac419a3-c8", "ovs_interfaceid": "fac419a3-c8d4-48b4-8dde-b6a64593a275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 764.744232] env[61995]: DEBUG oslo_concurrency.lockutils [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] Acquired lock "refresh_cache-2da17a43-2260-4a82-9ed3-eafd25bb9a6a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.744431] env[61995]: DEBUG nova.network.neutron [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Refreshing network info cache for port fac419a3-c8d4-48b4-8dde-b6a64593a275 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 764.745622] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:68:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c297fe21-cd0b-4226-813b-a65d2358d034', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fac419a3-c8d4-48b4-8dde-b6a64593a275', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 764.753911] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Creating folder: Project (556cd4687d78408bbd5e53e3aec7eecc). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 764.759023] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7e4ac27-c298-4243-b11c-cc707a1eec70 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.768485] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Created folder: Project (556cd4687d78408bbd5e53e3aec7eecc) in parent group-v185203. [ 764.768703] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Creating folder: Instances. Parent ref: group-v185268. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 764.768986] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-735fc243-33a2-4a8d-8a8d-1b53a29e07c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.777502] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Created folder: Instances in parent group-v185268. [ 764.777780] env[61995]: DEBUG oslo.service.loopingcall [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.777981] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 764.778226] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4432b98-f624-4195-b240-6d1355442a81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.800961] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 764.800961] env[61995]: value = "task-794410" [ 764.800961] env[61995]: _type = "Task" [ 764.800961] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.808968] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794410, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.824105] env[61995]: DEBUG nova.scheduler.client.report [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.828576] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.052506] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7626dad-2cc3-4493-a981-66c82f25ce17 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.089s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.070206] env[61995]: DEBUG nova.network.neutron [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Updated VIF entry in instance network info cache for port fac419a3-c8d4-48b4-8dde-b6a64593a275. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 765.070645] env[61995]: DEBUG nova.network.neutron [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Updating instance_info_cache with network_info: [{"id": "fac419a3-c8d4-48b4-8dde-b6a64593a275", "address": "fa:16:3e:dc:68:18", "network": {"id": "b9bf6713-8a48-4ec9-ab6f-4972939c5daf", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-889552615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "556cd4687d78408bbd5e53e3aec7eecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfac419a3-c8", "ovs_interfaceid": "fac419a3-c8d4-48b4-8dde-b6a64593a275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.089948] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794406, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.173241] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "0348098c-fd4c-41f8-a246-b9ba87a75674" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.173773] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.230998] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260389} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.230998] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.230998] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94eed030-aa8a-4039-9652-0018438b8849 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.263840] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 4da426a1-7af9-4e4e-82ed-c45687a19b86/4da426a1-7af9-4e4e-82ed-c45687a19b86.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.264438] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-833d10d6-dd61-4655-b2a9-44d7c1dc2909 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.284788] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 765.284788] env[61995]: value = "task-794411" [ 765.284788] env[61995]: _type = "Task" [ 765.284788] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.293750] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794411, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.311398] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794410, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.330326] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.555s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.330856] env[61995]: DEBUG nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 765.333485] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.911s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.333698] env[61995]: DEBUG nova.objects.instance [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lazy-loading 'resources' on Instance uuid 19fd0358-32e4-434f-b0b3-c8cc6c07ba51 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 765.555917] env[61995]: DEBUG nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.573521] env[61995]: DEBUG oslo_concurrency.lockutils [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] Releasing lock "refresh_cache-2da17a43-2260-4a82-9ed3-eafd25bb9a6a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.573521] env[61995]: DEBUG nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Received event network-vif-deleted-a1cabe4c-61ba-460d-b384-421305897add {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.573663] env[61995]: INFO nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Neutron deleted interface a1cabe4c-61ba-460d-b384-421305897add; detaching it from the instance and deleting it from the info cache [ 765.573772] env[61995]: DEBUG nova.network.neutron [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.584050] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794406, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.795102] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794411, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.812205] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794410, 'name': CreateVM_Task, 'duration_secs': 0.558644} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.812205] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 765.812912] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.813401] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.813837] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 765.814755] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e71be765-12be-44b0-a376-972dcbe26a94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.821516] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 765.821516] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527cf838-bb23-381d-f249-20a8d2cedd5a" [ 765.821516] env[61995]: _type = "Task" [ 765.821516] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.831545] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527cf838-bb23-381d-f249-20a8d2cedd5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.836875] env[61995]: DEBUG nova.compute.utils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.841877] env[61995]: DEBUG nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.842078] env[61995]: DEBUG nova.network.neutron [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 765.912923] env[61995]: DEBUG nova.policy [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7e1129f16ae479f9f77d03439cd1211', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb4715780da643999551e223caa16065', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 766.079379] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.082340] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1659d2a8-d9f1-4364-89cd-949ddc91aca4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.089526] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794406, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.093167] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048043dd-cd43-45d0-9d21-372ed1168cf9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.125246] env[61995]: DEBUG nova.compute.manager [req-b0a0f231-fcfe-4e20-a6a5-cebc918a0ed0 req-de8db16d-ad5c-47db-9ce1-8a8bb5f73b9f service nova] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Detach interface failed, port_id=a1cabe4c-61ba-460d-b384-421305897add, reason: Instance 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 766.219357] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26168db2-b6f9-4a53-9c8e-2ce6a9d6da39 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.228019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57af8ad-85bc-4849-9225-355ee855fb82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.256534] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30384d37-3019-4c1f-8859-df469b5a25b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.266157] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6833e3-ed66-496e-9179-83a470d6b85f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.279519] env[61995]: DEBUG nova.compute.provider_tree [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.281287] env[61995]: DEBUG nova.network.neutron [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Successfully created port: 9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.295141] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794411, 'name': ReconfigVM_Task, 'duration_secs': 0.882814} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.295409] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 4da426a1-7af9-4e4e-82ed-c45687a19b86/4da426a1-7af9-4e4e-82ed-c45687a19b86.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.296066] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00924e09-5e85-4333-98cb-b6cfc1db45f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.301818] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 766.301818] env[61995]: value = "task-794412" [ 766.301818] env[61995]: _type = "Task" [ 766.301818] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.309908] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794412, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.335386] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527cf838-bb23-381d-f249-20a8d2cedd5a, 'name': SearchDatastore_Task, 'duration_secs': 0.033487} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.335727] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.335993] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.336388] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.336571] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.336780] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.337061] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-791a2e06-f241-4f0a-a022-4c8ba046177b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.342244] env[61995]: DEBUG nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.348217] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.348217] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 766.348217] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c33eac15-4ac7-48de-bd70-7d195de13f58 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.352475] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 766.352475] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ec0161-457c-2525-86a5-f332c0e30c55" [ 766.352475] env[61995]: _type = "Task" [ 766.352475] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.360917] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ec0161-457c-2525-86a5-f332c0e30c55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.391143] env[61995]: DEBUG nova.compute.manager [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 766.585829] env[61995]: DEBUG oslo_vmware.api [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794406, 'name': PowerOnVM_Task, 'duration_secs': 2.12095} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.586191] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 766.586436] env[61995]: DEBUG nova.compute.manager [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 766.587243] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8521da21-d1df-4ffb-9727-84b53e479e7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.784828] env[61995]: DEBUG nova.scheduler.client.report [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.814021] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794412, 'name': Rename_Task, 'duration_secs': 0.159689} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.814021] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 766.814021] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1651d4a1-fe57-4049-80fa-f1ca6003cb6e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.818971] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 766.818971] env[61995]: value = "task-794413" [ 766.818971] env[61995]: _type = "Task" [ 766.818971] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.826369] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794413, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.862846] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ec0161-457c-2525-86a5-f332c0e30c55, 'name': SearchDatastore_Task, 'duration_secs': 0.011797} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.864072] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4c30cd8-7665-43d5-abf8-1624b2f53778 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.868836] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 766.868836] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526e2c41-59da-1cf7-5629-23fc6c926303" [ 766.868836] env[61995]: _type = "Task" [ 766.868836] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.876364] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526e2c41-59da-1cf7-5629-23fc6c926303, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.906680] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.103160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.289885] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.956s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.292302] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.715s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.292530] env[61995]: DEBUG nova.objects.instance [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lazy-loading 'resources' on Instance uuid c350194c-ed3e-4996-871a-70b4cb99c7db {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.310742] env[61995]: INFO nova.scheduler.client.report [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Deleted allocations for instance 19fd0358-32e4-434f-b0b3-c8cc6c07ba51 [ 767.331184] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794413, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.354326] env[61995]: DEBUG nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.380694] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526e2c41-59da-1cf7-5629-23fc6c926303, 'name': SearchDatastore_Task, 'duration_secs': 0.011338} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.382870] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.383170] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2da17a43-2260-4a82-9ed3-eafd25bb9a6a/2da17a43-2260-4a82-9ed3-eafd25bb9a6a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 767.383669] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02a4f664-a2b2-4f5f-8bb3-c5281044bc33 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.390066] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.390322] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.390606] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.390689] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.390845] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.390997] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.391325] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.391519] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.391688] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.391854] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.392510] env[61995]: DEBUG nova.virt.hardware [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.392857] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82af27ac-dabf-4427-a0eb-9a6afb4f304a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.397147] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 767.397147] env[61995]: value = "task-794414" [ 767.397147] env[61995]: _type = "Task" [ 767.397147] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.405299] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb1bea6-480c-4ce4-b82f-13373205b222 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.413318] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794414, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.777297] env[61995]: DEBUG nova.compute.manager [req-e2f6995b-fa62-4ab1-88f2-4c0b88939a1e req-9189a79f-b6f5-4b70-aad3-6b3102c43408 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Received event network-vif-plugged-9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 767.777597] env[61995]: DEBUG oslo_concurrency.lockutils [req-e2f6995b-fa62-4ab1-88f2-4c0b88939a1e req-9189a79f-b6f5-4b70-aad3-6b3102c43408 service nova] Acquiring lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.777892] env[61995]: DEBUG oslo_concurrency.lockutils [req-e2f6995b-fa62-4ab1-88f2-4c0b88939a1e req-9189a79f-b6f5-4b70-aad3-6b3102c43408 service nova] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.778424] env[61995]: DEBUG oslo_concurrency.lockutils [req-e2f6995b-fa62-4ab1-88f2-4c0b88939a1e req-9189a79f-b6f5-4b70-aad3-6b3102c43408 service nova] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.778726] env[61995]: DEBUG nova.compute.manager [req-e2f6995b-fa62-4ab1-88f2-4c0b88939a1e req-9189a79f-b6f5-4b70-aad3-6b3102c43408 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] No waiting events found dispatching network-vif-plugged-9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 767.779035] env[61995]: WARNING nova.compute.manager [req-e2f6995b-fa62-4ab1-88f2-4c0b88939a1e req-9189a79f-b6f5-4b70-aad3-6b3102c43408 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Received unexpected event network-vif-plugged-9551a93f-ec41-48c7-b25a-aa038eca7341 for instance with vm_state building and task_state spawning. [ 767.819472] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a368ee88-0034-4970-9e55-97b2289962d4 tempest-AttachInterfacesV270Test-122437301 tempest-AttachInterfacesV270Test-122437301-project-member] Lock "19fd0358-32e4-434f-b0b3-c8cc6c07ba51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.351s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.845107] env[61995]: DEBUG oslo_vmware.api [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794413, 'name': PowerOnVM_Task, 'duration_secs': 0.574504} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.845371] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 767.845562] env[61995]: INFO nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Took 8.45 seconds to spawn the instance on the hypervisor. [ 767.845733] env[61995]: DEBUG nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.846607] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3dad2a-65b3-488b-bef9-cbe191de1195 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.912700] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794414, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.927438] env[61995]: DEBUG nova.network.neutron [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Successfully updated port: 9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 768.235493] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600d01e9-ab94-4655-83f7-fbf611936f43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.243398] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7991092-0795-42ab-aacd-e62bbeae92ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.273993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.274279] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.274489] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.274673] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.274942] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.277730] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f222b348-a1e0-4fef-9658-57ae86d37044 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.280492] env[61995]: INFO nova.compute.manager [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Terminating instance [ 768.282370] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "refresh_cache-ecb72668-caa4-4117-9f8d-e0c6e4831f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.282518] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "refresh_cache-ecb72668-caa4-4117-9f8d-e0c6e4831f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.282677] env[61995]: DEBUG nova.network.neutron [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.286778] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a6a626-06e2-4959-abea-0353ce0f7990 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.300708] env[61995]: DEBUG nova.compute.provider_tree [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.366828] env[61995]: INFO nova.compute.manager [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Took 41.35 seconds to build instance. [ 768.410848] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794414, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.872864} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.410848] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2da17a43-2260-4a82-9ed3-eafd25bb9a6a/2da17a43-2260-4a82-9ed3-eafd25bb9a6a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 768.410848] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 768.410848] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78543cc1-d513-4d4b-a7ff-6488ac8f96c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.417018] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 768.417018] env[61995]: value = "task-794415" [ 768.417018] env[61995]: _type = "Task" [ 768.417018] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.422682] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.429704] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.430087] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquired lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.430400] env[61995]: DEBUG nova.network.neutron [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.803987] env[61995]: DEBUG nova.scheduler.client.report [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.811434] env[61995]: DEBUG nova.network.neutron [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.870566] env[61995]: DEBUG oslo_concurrency.lockutils [None req-28ade485-f7f1-4168-b1e6-b72486f0f14e tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.402s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.887832] env[61995]: DEBUG nova.network.neutron [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.925681] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067599} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.925954] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.926801] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033b36c6-5c89-4c79-b908-fa264f3bcee6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.956787] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 2da17a43-2260-4a82-9ed3-eafd25bb9a6a/2da17a43-2260-4a82-9ed3-eafd25bb9a6a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.957248] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-682dca73-c6f9-458e-9e79-d1011e7dcb87 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.978644] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 768.978644] env[61995]: value = "task-794416" [ 768.978644] env[61995]: _type = "Task" [ 768.978644] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.989910] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794416, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.021422] env[61995]: DEBUG nova.network.neutron [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.196385] env[61995]: DEBUG nova.network.neutron [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updating instance_info_cache with network_info: [{"id": "9551a93f-ec41-48c7-b25a-aa038eca7341", "address": "fa:16:3e:d2:6c:28", "network": {"id": "616dc7e4-858b-4038-bfd9-30dc98a41f73", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-733696429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb4715780da643999551e223caa16065", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551a93f-ec", "ovs_interfaceid": "9551a93f-ec41-48c7-b25a-aa038eca7341", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.315534] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.023s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.317909] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.307s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.318237] env[61995]: DEBUG nova.objects.instance [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lazy-loading 'resources' on Instance uuid 45fdae21-bec2-438a-8f2a-c23b0e428057 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 769.343253] env[61995]: INFO nova.scheduler.client.report [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Deleted allocations for instance c350194c-ed3e-4996-871a-70b4cb99c7db [ 769.375508] env[61995]: DEBUG nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.396023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "refresh_cache-ecb72668-caa4-4117-9f8d-e0c6e4831f3d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.396023] env[61995]: DEBUG nova.compute.manager [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.396023] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 769.397609] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3041f4c5-d824-41c9-94ac-5c597ac77c7b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.409295] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 769.412571] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec868ea6-5366-48e8-87e6-7b4395f373e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.416187] env[61995]: DEBUG oslo_vmware.api [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 769.416187] env[61995]: value = "task-794417" [ 769.416187] env[61995]: _type = "Task" [ 769.416187] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.424472] env[61995]: DEBUG oslo_vmware.api [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.489447] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794416, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.699766] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Releasing lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.699922] env[61995]: DEBUG nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Instance network_info: |[{"id": "9551a93f-ec41-48c7-b25a-aa038eca7341", "address": "fa:16:3e:d2:6c:28", "network": {"id": "616dc7e4-858b-4038-bfd9-30dc98a41f73", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-733696429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb4715780da643999551e223caa16065", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551a93f-ec", "ovs_interfaceid": "9551a93f-ec41-48c7-b25a-aa038eca7341", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 769.700372] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:6c:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9551a93f-ec41-48c7-b25a-aa038eca7341', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.724287] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Creating folder: Project (bb4715780da643999551e223caa16065). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 769.725028] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e9e6ebd-b812-41eb-8fc1-bafbbb39dfde {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.735265] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Created folder: Project (bb4715780da643999551e223caa16065) in parent group-v185203. [ 769.735471] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Creating folder: Instances. Parent ref: group-v185271. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 769.735820] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3329fe42-f20b-46b9-861d-e221b7c828a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.744800] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Created folder: Instances in parent group-v185271. [ 769.745100] env[61995]: DEBUG oslo.service.loopingcall [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.749027] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 769.749027] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ffdeef81-00cf-4d9e-8a80-70c7bf5ff7e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.764655] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.764655] env[61995]: value = "task-794420" [ 769.764655] env[61995]: _type = "Task" [ 769.764655] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.772906] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794420, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.812850] env[61995]: DEBUG nova.compute.manager [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Received event network-changed-9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 769.813136] env[61995]: DEBUG nova.compute.manager [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Refreshing instance network info cache due to event network-changed-9551a93f-ec41-48c7-b25a-aa038eca7341. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 769.813282] env[61995]: DEBUG oslo_concurrency.lockutils [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] Acquiring lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.813426] env[61995]: DEBUG oslo_concurrency.lockutils [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] Acquired lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.813580] env[61995]: DEBUG nova.network.neutron [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Refreshing network info cache for port 9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.852607] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4e0b271c-89d8-4740-9986-583e294ecac2 tempest-ServersNegativeTestMultiTenantJSON-903261949 tempest-ServersNegativeTestMultiTenantJSON-903261949-project-member] Lock "c350194c-ed3e-4996-871a-70b4cb99c7db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.157s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.903047] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.932266] env[61995]: DEBUG oslo_vmware.api [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794417, 'name': PowerOffVM_Task, 'duration_secs': 0.502678} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.934946] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 769.935141] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 769.936450] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e6a23c1-63df-4aec-a922-81eefd602ea5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.967062] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 769.967299] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 769.967479] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleting the datastore file [datastore2] ecb72668-caa4-4117-9f8d-e0c6e4831f3d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 769.967732] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be5cd960-2698-4780-86a6-2cc91d095401 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.975714] env[61995]: DEBUG oslo_vmware.api [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 769.975714] env[61995]: value = "task-794422" [ 769.975714] env[61995]: _type = "Task" [ 769.975714] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.988367] env[61995]: DEBUG oslo_vmware.api [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.994106] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794416, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.054962] env[61995]: DEBUG nova.compute.manager [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.055869] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4029d4d-04b7-4e62-ad3c-c7e8bab047a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.254609] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916e80ef-5e33-47a2-8d42-81201be485ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.263449] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294b7719-4973-4f39-a416-e244853858ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.273954] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794420, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.300568] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36974a6f-2593-4a55-888c-4372345eb3bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.307916] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f837276-43bc-44df-be41-9404be4f4cce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.323511] env[61995]: DEBUG nova.compute.provider_tree [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.491700] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794416, 'name': ReconfigVM_Task, 'duration_secs': 1.325024} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.494583] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 2da17a43-2260-4a82-9ed3-eafd25bb9a6a/2da17a43-2260-4a82-9ed3-eafd25bb9a6a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.495245] env[61995]: DEBUG oslo_vmware.api [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174303} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.495447] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-640d229e-91b3-42ec-a0a9-1649b61cd66e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.496964] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 770.497149] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 770.497325] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 770.497524] env[61995]: INFO nova.compute.manager [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 770.497710] env[61995]: DEBUG oslo.service.loopingcall [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.497889] env[61995]: DEBUG nova.compute.manager [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.498017] env[61995]: DEBUG nova.network.neutron [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.505064] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 770.505064] env[61995]: value = "task-794423" [ 770.505064] env[61995]: _type = "Task" [ 770.505064] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.514307] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794423, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.516463] env[61995]: DEBUG nova.network.neutron [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.569770] env[61995]: INFO nova.compute.manager [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] instance snapshotting [ 770.572516] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994bad2b-b801-4955-96e0-331a129a0a84 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.596766] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26eedce7-966b-466e-bde8-899578ad1ec7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.684312] env[61995]: DEBUG nova.network.neutron [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updated VIF entry in instance network info cache for port 9551a93f-ec41-48c7-b25a-aa038eca7341. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.684649] env[61995]: DEBUG nova.network.neutron [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updating instance_info_cache with network_info: [{"id": "9551a93f-ec41-48c7-b25a-aa038eca7341", "address": "fa:16:3e:d2:6c:28", "network": {"id": "616dc7e4-858b-4038-bfd9-30dc98a41f73", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-733696429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb4715780da643999551e223caa16065", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551a93f-ec", "ovs_interfaceid": "9551a93f-ec41-48c7-b25a-aa038eca7341", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.778245] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794420, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.831579] env[61995]: DEBUG nova.scheduler.client.report [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.017795] env[61995]: DEBUG nova.network.neutron [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.020450] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794423, 'name': Rename_Task, 'duration_secs': 0.183605} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.020840] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 771.021304] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42f7d1e4-2055-4a8f-9789-9ee24b04f6f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.036741] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 771.036741] env[61995]: value = "task-794424" [ 771.036741] env[61995]: _type = "Task" [ 771.036741] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.044361] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794424, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.109844] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 771.114742] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4528dc9d-ac9a-4916-8181-05628accca77 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.122183] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 771.122183] env[61995]: value = "task-794425" [ 771.122183] env[61995]: _type = "Task" [ 771.122183] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.130346] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794425, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.187720] env[61995]: DEBUG oslo_concurrency.lockutils [req-f9034d24-47ff-4828-b80d-9e331f0d1a6b req-b52648cd-9e90-4a4a-852e-3b8a8d4747c9 service nova] Releasing lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.277661] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794420, 'name': CreateVM_Task, 'duration_secs': 1.451022} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.277875] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 771.282715] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.282930] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.283270] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.283557] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83e2154b-9d50-4a7d-8594-259afdafa8fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.290316] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 771.290316] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52cc7fcd-a3a0-9678-10d7-da522fc54410" [ 771.290316] env[61995]: _type = "Task" [ 771.290316] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.300930] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52cc7fcd-a3a0-9678-10d7-da522fc54410, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.336283] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.018s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.338176] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.454s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.339844] env[61995]: INFO nova.compute.claims [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.365279] env[61995]: INFO nova.scheduler.client.report [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted allocations for instance 45fdae21-bec2-438a-8f2a-c23b0e428057 [ 771.527670] env[61995]: INFO nova.compute.manager [-] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Took 1.03 seconds to deallocate network for instance. [ 771.544953] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794424, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.633450] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794425, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.802863] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52cc7fcd-a3a0-9678-10d7-da522fc54410, 'name': SearchDatastore_Task, 'duration_secs': 0.009509} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.803382] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.803844] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 771.805017] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.805017] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.805017] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 771.805017] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea918ae7-0dfd-4e65-a51b-b278c6dbce21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.816329] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 771.816893] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 771.817739] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84a8283c-b419-4030-940d-8a51dfdad8fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.825659] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 771.825659] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523f0c45-698a-5476-f2b3-cc27b2ecebee" [ 771.825659] env[61995]: _type = "Task" [ 771.825659] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.836456] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523f0c45-698a-5476-f2b3-cc27b2ecebee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.874795] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fc9a7ad0-97ec-445d-963c-dfbbd56fca88 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "45fdae21-bec2-438a-8f2a-c23b0e428057" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.876s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.040314] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.047510] env[61995]: DEBUG oslo_vmware.api [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794424, 'name': PowerOnVM_Task, 'duration_secs': 0.90118} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.047835] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 772.048115] env[61995]: INFO nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Took 10.19 seconds to spawn the instance on the hypervisor. [ 772.048358] env[61995]: DEBUG nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.049409] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3552eb25-0498-48a6-bddc-9cf06b79516d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.137219] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794425, 'name': CreateSnapshot_Task, 'duration_secs': 0.839581} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.137988] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 772.138567] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94ac4cf-1795-448a-91ec-91b4f15474a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.340820] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523f0c45-698a-5476-f2b3-cc27b2ecebee, 'name': SearchDatastore_Task, 'duration_secs': 0.010098} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.343989] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f15ab87-c063-439c-a41b-1e9f5d4e9e86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.349872] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 772.349872] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5296141d-8b8d-0abe-fabd-023cf93eddc1" [ 772.349872] env[61995]: _type = "Task" [ 772.349872] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.362764] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5296141d-8b8d-0abe-fabd-023cf93eddc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.572569] env[61995]: INFO nova.compute.manager [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Took 45.28 seconds to build instance. [ 772.660787] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 772.661859] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f05531b5-8854-413a-87cd-57c5f2b40902 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.677421] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 772.677421] env[61995]: value = "task-794426" [ 772.677421] env[61995]: _type = "Task" [ 772.677421] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.688632] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794426, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.757018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d84e7b2-15ec-45a9-b498-c3bbaa9e8901 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.765784] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d4bf41-ea02-44d5-b131-37f7fd99fe8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.804459] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae5c4ee-7679-4e79-b027-3ba019787a8d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.813164] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09010b05-fd0c-419e-97af-c1bd8d3ee466 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.830163] env[61995]: DEBUG nova.compute.provider_tree [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.865596] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5296141d-8b8d-0abe-fabd-023cf93eddc1, 'name': SearchDatastore_Task, 'duration_secs': 0.028485} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.865868] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.866192] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] c51184b6-d2c7-49ff-b061-ab0f9a69cdbf/c51184b6-d2c7-49ff-b061-ab0f9a69cdbf.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 772.866666] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b85cd3d-0812-49c6-b26a-fd9f7092670c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.874345] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 772.874345] env[61995]: value = "task-794427" [ 772.874345] env[61995]: _type = "Task" [ 772.874345] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.884634] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794427, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.076766] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1024f15d-55f2-4306-b165-393d48a2e687 tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.705s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.194508] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794426, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.335162] env[61995]: DEBUG nova.scheduler.client.report [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.386494] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794427, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.582953] env[61995]: DEBUG nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.692826] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794426, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.724044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.724044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.724044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.724044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.724044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.728601] env[61995]: INFO nova.compute.manager [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Terminating instance [ 773.729773] env[61995]: DEBUG nova.compute.manager [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 773.729952] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 773.730811] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6837f6e7-d278-4caa-92cd-8e2934a20f8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.741017] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 773.741017] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dea3d1ea-0940-42c8-938c-19e0f931dafa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.745579] env[61995]: DEBUG oslo_vmware.api [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 773.745579] env[61995]: value = "task-794428" [ 773.745579] env[61995]: _type = "Task" [ 773.745579] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.753553] env[61995]: DEBUG oslo_vmware.api [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794428, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.843027] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.843027] env[61995]: DEBUG nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.846051] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.342s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.847778] env[61995]: INFO nova.compute.claims [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.891990] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794427, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532927} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.892554] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] c51184b6-d2c7-49ff-b061-ab0f9a69cdbf/c51184b6-d2c7-49ff-b061-ab0f9a69cdbf.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 773.892554] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 773.892817] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30d595d0-f869-44c0-870a-05f9b7981806 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.902495] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 773.902495] env[61995]: value = "task-794429" [ 773.902495] env[61995]: _type = "Task" [ 773.902495] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.916869] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794429, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.104033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.197018] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794426, 'name': CloneVM_Task} progress is 95%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.263332] env[61995]: DEBUG oslo_vmware.api [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794428, 'name': PowerOffVM_Task, 'duration_secs': 0.178963} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.263332] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 774.263332] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 774.264763] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b519c30-2947-46c7-9a3d-bc13bbd2f09e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.333660] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 774.333901] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 774.334100] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Deleting the datastore file [datastore2] 2da17a43-2260-4a82-9ed3-eafd25bb9a6a {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 774.334370] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9dfe9601-6bd0-4673-8fa4-98ce7518a5de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.341489] env[61995]: DEBUG oslo_vmware.api [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for the task: (returnval){ [ 774.341489] env[61995]: value = "task-794431" [ 774.341489] env[61995]: _type = "Task" [ 774.341489] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.348691] env[61995]: DEBUG nova.compute.utils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.353565] env[61995]: DEBUG nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.353800] env[61995]: DEBUG nova.network.neutron [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 774.358273] env[61995]: DEBUG oslo_vmware.api [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.409429] env[61995]: DEBUG nova.policy [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '125f1d35f68844da880383e33fef0644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99211c1a5bd34212848f45929771eb71', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 774.415783] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794429, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074568} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.416986] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.416986] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d472136-b520-4952-b6da-450a7945a3d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.439067] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] c51184b6-d2c7-49ff-b061-ab0f9a69cdbf/c51184b6-d2c7-49ff-b061-ab0f9a69cdbf.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.439714] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0760028b-fba4-4615-a865-ccfb87483a22 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.461380] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 774.461380] env[61995]: value = "task-794432" [ 774.461380] env[61995]: _type = "Task" [ 774.461380] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.468291] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794432, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.636293] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "7524c912-4d73-462a-949f-3b4c85874094" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.637025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "7524c912-4d73-462a-949f-3b4c85874094" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.665232] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.665467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.696155] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794426, 'name': CloneVM_Task, 'duration_secs': 2.014991} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.696276] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Created linked-clone VM from snapshot [ 774.699817] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76233e5-cab1-47cb-96c7-c148b9dd9676 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.702803] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "af7163e8-4e77-4336-80e9-0623e29ed8c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.703256] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.708979] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Uploading image c24b00a5-c673-4ee2-8bcb-389e5191f7fe {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 774.746385] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 774.746385] env[61995]: value = "vm-185275" [ 774.746385] env[61995]: _type = "VirtualMachine" [ 774.746385] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 774.746649] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ef7f9b16-74a6-4a0a-a785-f9d5b30a2621 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.756474] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lease: (returnval){ [ 774.756474] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d27c26-9d79-498c-8f91-17c31ab4a9ac" [ 774.756474] env[61995]: _type = "HttpNfcLease" [ 774.756474] env[61995]: } obtained for exporting VM: (result){ [ 774.756474] env[61995]: value = "vm-185275" [ 774.756474] env[61995]: _type = "VirtualMachine" [ 774.756474] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 774.756866] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the lease: (returnval){ [ 774.756866] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d27c26-9d79-498c-8f91-17c31ab4a9ac" [ 774.756866] env[61995]: _type = "HttpNfcLease" [ 774.756866] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 774.763573] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 774.763573] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d27c26-9d79-498c-8f91-17c31ab4a9ac" [ 774.763573] env[61995]: _type = "HttpNfcLease" [ 774.763573] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 774.773701] env[61995]: DEBUG nova.network.neutron [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Successfully created port: b2df31cb-ec39-4c6d-8011-9ab55731a25f {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.851926] env[61995]: DEBUG nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.854366] env[61995]: DEBUG oslo_vmware.api [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.981906] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.266297] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 775.266297] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d27c26-9d79-498c-8f91-17c31ab4a9ac" [ 775.266297] env[61995]: _type = "HttpNfcLease" [ 775.266297] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 775.266601] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 775.266601] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d27c26-9d79-498c-8f91-17c31ab4a9ac" [ 775.266601] env[61995]: _type = "HttpNfcLease" [ 775.266601] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 775.268398] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb350ccb-ac01-4566-877e-0913bec6a569 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.273019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00cbe87-65bb-4d46-a6a5-42305ffd1882 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.280649] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a2b5-9368-268b-7c7d-bc956f781b08/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 775.280817] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a2b5-9368-268b-7c7d-bc956f781b08/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 775.285188] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de9f29e-e673-400b-8784-7212768a5c55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.383730] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc7ee4e-0bd7-47cc-b09d-930091687543 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.386636] env[61995]: DEBUG oslo_vmware.api [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Task: {'id': task-794431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.654128} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.387211] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 775.387763] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 775.387763] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 775.387763] env[61995]: INFO nova.compute.manager [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Took 1.66 seconds to destroy the instance on the hypervisor. [ 775.391026] env[61995]: DEBUG oslo.service.loopingcall [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.391026] env[61995]: DEBUG nova.compute.manager [-] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.391026] env[61995]: DEBUG nova.network.neutron [-] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 775.393733] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bd9dea-f7e0-4cc8-8685-49e167f52d17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.408433] env[61995]: DEBUG nova.compute.provider_tree [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.448051] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3efafdfb-2881-454c-b803-89832fe38b27 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.473199] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794432, 'name': ReconfigVM_Task, 'duration_secs': 0.78341} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.476014] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Reconfigured VM instance instance-0000002c to attach disk [datastore2] c51184b6-d2c7-49ff-b061-ab0f9a69cdbf/c51184b6-d2c7-49ff-b061-ab0f9a69cdbf.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.478106] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40940f75-584d-4236-bf62-dc8af20c0362 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.487794] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 775.487794] env[61995]: value = "task-794434" [ 775.487794] env[61995]: _type = "Task" [ 775.487794] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.496331] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794434, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.654927] env[61995]: DEBUG nova.compute.manager [req-59b32c56-9395-4d3d-a597-cb535b8dad03 req-d364a614-4633-4438-ba8a-d172af4d1e35 service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Received event network-vif-deleted-fac419a3-c8d4-48b4-8dde-b6a64593a275 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.655196] env[61995]: INFO nova.compute.manager [req-59b32c56-9395-4d3d-a597-cb535b8dad03 req-d364a614-4633-4438-ba8a-d172af4d1e35 service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Neutron deleted interface fac419a3-c8d4-48b4-8dde-b6a64593a275; detaching it from the instance and deleting it from the info cache [ 775.655381] env[61995]: DEBUG nova.network.neutron [req-59b32c56-9395-4d3d-a597-cb535b8dad03 req-d364a614-4633-4438-ba8a-d172af4d1e35 service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.889831] env[61995]: DEBUG nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.909958] env[61995]: DEBUG nova.scheduler.client.report [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.922333] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.922333] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.922725] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.922725] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.922806] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.922931] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.923350] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.923545] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.923769] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.923986] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.924251] env[61995]: DEBUG nova.virt.hardware [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.925536] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06162858-af14-4d96-9f33-9f0461406c13 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.935701] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c201a8-f7b3-42c8-b66c-330d4c12c60e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.998920] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794434, 'name': Rename_Task, 'duration_secs': 0.233441} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.999332] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 775.999537] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41db07d2-c36c-4e0c-aaad-1433273d751c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.007163] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 776.007163] env[61995]: value = "task-794435" [ 776.007163] env[61995]: _type = "Task" [ 776.007163] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.015649] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794435, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.134056] env[61995]: DEBUG nova.network.neutron [-] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.160464] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e2350c3-06a0-4022-a041-7d0344697acf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.169923] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f4c335-109c-40fb-9efb-205de4cb8cfe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.198439] env[61995]: DEBUG nova.compute.manager [req-59b32c56-9395-4d3d-a597-cb535b8dad03 req-d364a614-4633-4438-ba8a-d172af4d1e35 service nova] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Detach interface failed, port_id=fac419a3-c8d4-48b4-8dde-b6a64593a275, reason: Instance 2da17a43-2260-4a82-9ed3-eafd25bb9a6a could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 776.281622] env[61995]: DEBUG nova.compute.manager [req-fd6ec995-72b0-4461-9531-8e504f922bce req-dc393485-148a-432a-a6e5-bbea42216baf service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Received event network-vif-plugged-b2df31cb-ec39-4c6d-8011-9ab55731a25f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 776.281961] env[61995]: DEBUG oslo_concurrency.lockutils [req-fd6ec995-72b0-4461-9531-8e504f922bce req-dc393485-148a-432a-a6e5-bbea42216baf service nova] Acquiring lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.282292] env[61995]: DEBUG oslo_concurrency.lockutils [req-fd6ec995-72b0-4461-9531-8e504f922bce req-dc393485-148a-432a-a6e5-bbea42216baf service nova] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.282560] env[61995]: DEBUG oslo_concurrency.lockutils [req-fd6ec995-72b0-4461-9531-8e504f922bce req-dc393485-148a-432a-a6e5-bbea42216baf service nova] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.282830] env[61995]: DEBUG nova.compute.manager [req-fd6ec995-72b0-4461-9531-8e504f922bce req-dc393485-148a-432a-a6e5-bbea42216baf service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] No waiting events found dispatching network-vif-plugged-b2df31cb-ec39-4c6d-8011-9ab55731a25f {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 776.283121] env[61995]: WARNING nova.compute.manager [req-fd6ec995-72b0-4461-9531-8e504f922bce req-dc393485-148a-432a-a6e5-bbea42216baf service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Received unexpected event network-vif-plugged-b2df31cb-ec39-4c6d-8011-9ab55731a25f for instance with vm_state building and task_state spawning. [ 776.415078] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.416636] env[61995]: DEBUG nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.421518] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.212s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.421813] env[61995]: DEBUG nova.objects.instance [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lazy-loading 'resources' on Instance uuid a04fbe7e-e53e-43ef-9a6f-ba3895ce7597 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 776.451385] env[61995]: DEBUG nova.network.neutron [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Successfully updated port: b2df31cb-ec39-4c6d-8011-9ab55731a25f {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.520035] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794435, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.638771] env[61995]: INFO nova.compute.manager [-] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Took 1.25 seconds to deallocate network for instance. [ 776.923840] env[61995]: DEBUG nova.compute.utils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.925361] env[61995]: DEBUG nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 776.925534] env[61995]: DEBUG nova.network.neutron [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 776.953676] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "refresh_cache-f6db353f-f7b5-4ba4-8f70-f3b65281da93" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.953840] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquired lock "refresh_cache-f6db353f-f7b5-4ba4-8f70-f3b65281da93" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.953990] env[61995]: DEBUG nova.network.neutron [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.987768] env[61995]: DEBUG nova.policy [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba6205c1911642f186c45b2b621aae92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1b27447bed34a35911f7585608107b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 777.020723] env[61995]: DEBUG oslo_vmware.api [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794435, 'name': PowerOnVM_Task, 'duration_secs': 0.783883} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.021034] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 777.021248] env[61995]: INFO nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Took 9.67 seconds to spawn the instance on the hypervisor. [ 777.021434] env[61995]: DEBUG nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 777.024901] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f3b73c-45b8-4191-b80b-7246e7367fe9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.145173] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.311159] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304569e8-afc4-4569-ad2f-fb8eb8b5b9d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.314913] env[61995]: DEBUG nova.network.neutron [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Successfully created port: d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.324362] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5938b987-628a-4509-a1bb-b23f4382f235 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.358482] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7100838d-a4e9-4b37-a052-6f2d2f56f216 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.366057] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab521d4-1891-4f8b-868d-1bb3531a7bcd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.382204] env[61995]: DEBUG nova.compute.provider_tree [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.431110] env[61995]: DEBUG nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.505042] env[61995]: DEBUG nova.network.neutron [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.545327] env[61995]: INFO nova.compute.manager [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Took 45.07 seconds to build instance. [ 777.703491] env[61995]: DEBUG nova.network.neutron [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Updating instance_info_cache with network_info: [{"id": "b2df31cb-ec39-4c6d-8011-9ab55731a25f", "address": "fa:16:3e:6a:41:6e", "network": {"id": "348741bc-5b79-4e0f-ae57-d81f6d3dfb8c", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1245066033-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99211c1a5bd34212848f45929771eb71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2df31cb-ec", "ovs_interfaceid": "b2df31cb-ec39-4c6d-8011-9ab55731a25f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.885024] env[61995]: DEBUG nova.scheduler.client.report [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.047836] env[61995]: DEBUG oslo_concurrency.lockutils [None req-461e2376-0560-4841-87db-498c3816f0cf tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.650s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.206734] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Releasing lock "refresh_cache-f6db353f-f7b5-4ba4-8f70-f3b65281da93" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.206734] env[61995]: DEBUG nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Instance network_info: |[{"id": "b2df31cb-ec39-4c6d-8011-9ab55731a25f", "address": "fa:16:3e:6a:41:6e", "network": {"id": "348741bc-5b79-4e0f-ae57-d81f6d3dfb8c", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1245066033-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99211c1a5bd34212848f45929771eb71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2df31cb-ec", "ovs_interfaceid": "b2df31cb-ec39-4c6d-8011-9ab55731a25f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 778.207043] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:41:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4813d311-2016-4f6a-a4b8-7613ab624fba', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2df31cb-ec39-4c6d-8011-9ab55731a25f', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.214538] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Creating folder: Project (99211c1a5bd34212848f45929771eb71). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.215350] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d6c4811-5efe-46e2-857d-02070e454b77 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.226902] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Created folder: Project (99211c1a5bd34212848f45929771eb71) in parent group-v185203. [ 778.227204] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Creating folder: Instances. Parent ref: group-v185276. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.227475] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23ecbb39-efa7-40e9-993b-8ae0cbca26dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.236947] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Created folder: Instances in parent group-v185276. [ 778.237276] env[61995]: DEBUG oslo.service.loopingcall [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.237520] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 778.237755] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da527e70-c13a-4705-9694-2100d005914b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.258019] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.258019] env[61995]: value = "task-794438" [ 778.258019] env[61995]: _type = "Task" [ 778.258019] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.266338] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794438, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.376400] env[61995]: DEBUG nova.compute.manager [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Received event network-changed-b2df31cb-ec39-4c6d-8011-9ab55731a25f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.376609] env[61995]: DEBUG nova.compute.manager [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Refreshing instance network info cache due to event network-changed-b2df31cb-ec39-4c6d-8011-9ab55731a25f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 778.376830] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] Acquiring lock "refresh_cache-f6db353f-f7b5-4ba4-8f70-f3b65281da93" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.377009] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] Acquired lock "refresh_cache-f6db353f-f7b5-4ba4-8f70-f3b65281da93" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.377213] env[61995]: DEBUG nova.network.neutron [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Refreshing network info cache for port b2df31cb-ec39-4c6d-8011-9ab55731a25f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.390517] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.969s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.392846] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.032s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.394439] env[61995]: INFO nova.compute.claims [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.414522] env[61995]: INFO nova.scheduler.client.report [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Deleted allocations for instance a04fbe7e-e53e-43ef-9a6f-ba3895ce7597 [ 778.445659] env[61995]: DEBUG nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.470591] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.470848] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.471041] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.471219] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.471370] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.471519] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.471730] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.471891] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.472082] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.472329] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.472481] env[61995]: DEBUG nova.virt.hardware [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.474662] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed386063-dde2-4cd0-956a-21fbd0db313e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.481975] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0879ca-d92f-4a22-8322-413931ec3619 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.549999] env[61995]: DEBUG nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 778.768689] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794438, 'name': CreateVM_Task, 'duration_secs': 0.419427} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.768885] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 778.769642] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.769823] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.770207] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.770490] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c901da94-2dac-4f38-a85d-2e03cee7e42f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.775392] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 778.775392] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d94422-68bb-38c4-0982-9bf015645e11" [ 778.775392] env[61995]: _type = "Task" [ 778.775392] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.779301] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.779427] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.779634] env[61995]: INFO nova.compute.manager [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Rebooting instance [ 778.784329] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d94422-68bb-38c4-0982-9bf015645e11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.882454] env[61995]: DEBUG nova.network.neutron [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Successfully updated port: d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 778.921589] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d4357e38-fe82-4268-88b6-260e841d8c10 tempest-VolumesAdminNegativeTest-1734529102 tempest-VolumesAdminNegativeTest-1734529102-project-member] Lock "a04fbe7e-e53e-43ef-9a6f-ba3895ce7597" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.684s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.074940] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.302844] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d94422-68bb-38c4-0982-9bf015645e11, 'name': SearchDatastore_Task, 'duration_secs': 0.019157} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.305080] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.305377] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.305618] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.305767] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.306438] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.306438] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fc448c2-bc01-4839-9459-4d15389ae81b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.320948] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.321166] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 779.321896] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9024a244-2fd8-4a3d-be5b-93b1fcac1fc0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.327668] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 779.327668] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]528097e6-b9ad-2e39-af3c-c2f81b438c7b" [ 779.327668] env[61995]: _type = "Task" [ 779.327668] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.328593] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.329635] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquired lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.329635] env[61995]: DEBUG nova.network.neutron [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.348317] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528097e6-b9ad-2e39-af3c-c2f81b438c7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.385422] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.385490] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquired lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.385692] env[61995]: DEBUG nova.network.neutron [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.499794] env[61995]: DEBUG nova.network.neutron [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Updated VIF entry in instance network info cache for port b2df31cb-ec39-4c6d-8011-9ab55731a25f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.500248] env[61995]: DEBUG nova.network.neutron [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Updating instance_info_cache with network_info: [{"id": "b2df31cb-ec39-4c6d-8011-9ab55731a25f", "address": "fa:16:3e:6a:41:6e", "network": {"id": "348741bc-5b79-4e0f-ae57-d81f6d3dfb8c", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1245066033-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99211c1a5bd34212848f45929771eb71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2df31cb-ec", "ovs_interfaceid": "b2df31cb-ec39-4c6d-8011-9ab55731a25f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.782322] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6e1b0b-bb8a-4056-a35e-6aab38cf2210 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.789998] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4a00fc-f4d5-4b2b-afa3-654060e312aa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.820984] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b7feeb-c465-4a4f-b4bf-27de6d062804 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.829275] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094ac118-f212-4f00-b97d-13a923980bbf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.845798] env[61995]: DEBUG nova.compute.provider_tree [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.859027] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528097e6-b9ad-2e39-af3c-c2f81b438c7b, 'name': SearchDatastore_Task, 'duration_secs': 0.024765} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.860966] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a78dcebd-0757-4a74-aadd-547086eb67bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.870037] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 779.870037] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]528e9eec-0378-875a-1552-2be340955a7b" [ 779.870037] env[61995]: _type = "Task" [ 779.870037] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.878516] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528e9eec-0378-875a-1552-2be340955a7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.940417] env[61995]: DEBUG nova.network.neutron [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.004703] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] Releasing lock "refresh_cache-f6db353f-f7b5-4ba4-8f70-f3b65281da93" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.005222] env[61995]: DEBUG nova.compute.manager [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Received event network-changed-9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.005222] env[61995]: DEBUG nova.compute.manager [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Refreshing instance network info cache due to event network-changed-9551a93f-ec41-48c7-b25a-aa038eca7341. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 780.005355] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] Acquiring lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.120710] env[61995]: DEBUG nova.network.neutron [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updating instance_info_cache with network_info: [{"id": "9551a93f-ec41-48c7-b25a-aa038eca7341", "address": "fa:16:3e:d2:6c:28", "network": {"id": "616dc7e4-858b-4038-bfd9-30dc98a41f73", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-733696429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb4715780da643999551e223caa16065", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551a93f-ec", "ovs_interfaceid": "9551a93f-ec41-48c7-b25a-aa038eca7341", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.158124] env[61995]: DEBUG nova.network.neutron [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.355179] env[61995]: DEBUG nova.scheduler.client.report [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.383924] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528e9eec-0378-875a-1552-2be340955a7b, 'name': SearchDatastore_Task, 'duration_secs': 0.013987} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.384234] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.384490] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] f6db353f-f7b5-4ba4-8f70-f3b65281da93/f6db353f-f7b5-4ba4-8f70-f3b65281da93.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 780.384747] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee220802-205f-4559-b5ba-1516521394bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.391688] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 780.391688] env[61995]: value = "task-794439" [ 780.391688] env[61995]: _type = "Task" [ 780.391688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.399897] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794439, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.623601] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Releasing lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.626262] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] Acquired lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.627150] env[61995]: DEBUG nova.network.neutron [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Refreshing network info cache for port 9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.632671] env[61995]: DEBUG nova.compute.manager [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.633581] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069add63-55e1-4fd4-b3e3-1d721e2e7a43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.661322] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Releasing lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.661671] env[61995]: DEBUG nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Instance network_info: |[{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 780.662123] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:57:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd33839ae-40ca-471b-92e3-eb282b920682', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd044777d-78bf-4b5b-a0dd-3b2dec017fbc', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.670352] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Creating folder: Project (b1b27447bed34a35911f7585608107b4). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 780.670628] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be8f1fc0-811a-4cbf-ae02-3f20e1153a2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.679286] env[61995]: DEBUG nova.compute.manager [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Received event network-vif-plugged-d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.679551] env[61995]: DEBUG oslo_concurrency.lockutils [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] Acquiring lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.679852] env[61995]: DEBUG oslo_concurrency.lockutils [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.680123] env[61995]: DEBUG oslo_concurrency.lockutils [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.680401] env[61995]: DEBUG nova.compute.manager [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] No waiting events found dispatching network-vif-plugged-d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 780.680664] env[61995]: WARNING nova.compute.manager [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Received unexpected event network-vif-plugged-d044777d-78bf-4b5b-a0dd-3b2dec017fbc for instance with vm_state building and task_state spawning. [ 780.680886] env[61995]: DEBUG nova.compute.manager [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Received event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.681150] env[61995]: DEBUG nova.compute.manager [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing instance network info cache due to event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 780.681558] env[61995]: DEBUG oslo_concurrency.lockutils [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] Acquiring lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.681761] env[61995]: DEBUG oslo_concurrency.lockutils [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] Acquired lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.681993] env[61995]: DEBUG nova.network.neutron [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.686080] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Created folder: Project (b1b27447bed34a35911f7585608107b4) in parent group-v185203. [ 780.686080] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Creating folder: Instances. Parent ref: group-v185279. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 780.686080] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ceb9eaf7-b191-4e4f-9abf-52eb844d147e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.698185] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Created folder: Instances in parent group-v185279. [ 780.698477] env[61995]: DEBUG oslo.service.loopingcall [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.699342] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 780.699479] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c22658e-b663-44f2-b48e-4654f7e4c88b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.721145] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.721145] env[61995]: value = "task-794442" [ 780.721145] env[61995]: _type = "Task" [ 780.721145] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.734675] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794442, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.864571] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.865147] env[61995]: DEBUG nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.869189] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.051s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.871511] env[61995]: INFO nova.compute.claims [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.904114] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794439, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.235068] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794442, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.378770] env[61995]: DEBUG nova.compute.utils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 781.382584] env[61995]: DEBUG nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 781.382980] env[61995]: DEBUG nova.network.neutron [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 781.412589] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794439, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.88205} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.412937] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] f6db353f-f7b5-4ba4-8f70-f3b65281da93/f6db353f-f7b5-4ba4-8f70-f3b65281da93.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 781.413296] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.413663] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d90f4a5-41e8-4905-a4ca-e3b6d7532bdb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.425738] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 781.425738] env[61995]: value = "task-794443" [ 781.425738] env[61995]: _type = "Task" [ 781.425738] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.435554] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794443, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.440167] env[61995]: DEBUG nova.network.neutron [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updated VIF entry in instance network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.440167] env[61995]: DEBUG nova.network.neutron [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.494426] env[61995]: DEBUG nova.policy [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '938522473ebd4eb18c2f7632c4ad73fc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '079a11ce311d40f9b8344fcce2e03683', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 781.609470] env[61995]: DEBUG nova.network.neutron [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updated VIF entry in instance network info cache for port 9551a93f-ec41-48c7-b25a-aa038eca7341. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.609910] env[61995]: DEBUG nova.network.neutron [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updating instance_info_cache with network_info: [{"id": "9551a93f-ec41-48c7-b25a-aa038eca7341", "address": "fa:16:3e:d2:6c:28", "network": {"id": "616dc7e4-858b-4038-bfd9-30dc98a41f73", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-733696429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb4715780da643999551e223caa16065", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551a93f-ec", "ovs_interfaceid": "9551a93f-ec41-48c7-b25a-aa038eca7341", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.651316] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7e139d-e263-4286-aa31-56141043b9e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.659449] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Doing hard reboot of VM {{(pid=61995) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 781.659789] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-b0cd76f5-e17d-4266-9043-9d7678111fcc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.666256] env[61995]: DEBUG oslo_vmware.api [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 781.666256] env[61995]: value = "task-794444" [ 781.666256] env[61995]: _type = "Task" [ 781.666256] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.675336] env[61995]: DEBUG oslo_vmware.api [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794444, 'name': ResetVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.732210] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794442, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.891439] env[61995]: DEBUG nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.903160] env[61995]: DEBUG nova.network.neutron [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Successfully created port: 9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.942677] env[61995]: DEBUG oslo_concurrency.lockutils [req-1401efed-ad12-4f9e-8ad7-3a4fcf8f2f14 req-f0fe99d0-d096-4d7d-a0fb-335e91fc87ac service nova] Releasing lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.943107] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794443, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080872} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.943358] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.944175] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242e6690-21d8-4946-9262-4769424b5216 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.972799] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] f6db353f-f7b5-4ba4-8f70-f3b65281da93/f6db353f-f7b5-4ba4-8f70-f3b65281da93.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.976678] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c59aae7-1ac8-4654-b72d-c91d58df39b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.002480] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 782.002480] env[61995]: value = "task-794445" [ 782.002480] env[61995]: _type = "Task" [ 782.002480] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.011961] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794445, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.112906] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0c1ed93-f6dd-47cf-950d-158a598cc395 req-919867d9-bd66-4249-8eed-20ac9f447f53 service nova] Releasing lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.180366] env[61995]: DEBUG oslo_vmware.api [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794444, 'name': ResetVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.232969] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794442, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.364801] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c432dea0-5c0b-46ce-aa01-76d0a8b480c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.373363] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e407df13-7013-4733-a2d8-fc5ca25b77ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.417834] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4020d7c2-c747-4c7a-8e4f-25eea58f8f48 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.426028] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361587c2-9218-458e-b610-2257427765b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.440968] env[61995]: DEBUG nova.compute.provider_tree [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.512583] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794445, 'name': ReconfigVM_Task, 'duration_secs': 0.333062} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.512883] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Reconfigured VM instance instance-0000002d to attach disk [datastore2] f6db353f-f7b5-4ba4-8f70-f3b65281da93/f6db353f-f7b5-4ba4-8f70-f3b65281da93.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.513542] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f3a465e-b8cd-4374-9669-966fd0ae9c1f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.520594] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 782.520594] env[61995]: value = "task-794446" [ 782.520594] env[61995]: _type = "Task" [ 782.520594] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.528755] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794446, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.677055] env[61995]: DEBUG oslo_vmware.api [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794444, 'name': ResetVM_Task, 'duration_secs': 0.906002} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.677342] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Did hard reboot of VM {{(pid=61995) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 782.677523] env[61995]: DEBUG nova.compute.manager [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.678381] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991b627f-4d0b-469d-965b-54b39fd120d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.735067] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794442, 'name': CreateVM_Task, 'duration_secs': 1.789111} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.735258] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.736053] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.736267] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.736650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.736975] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a98cce4-dae1-4f62-b39c-531e7d2025fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.741653] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 782.741653] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52154640-0ed6-7c4d-80f6-a0b0593fea13" [ 782.741653] env[61995]: _type = "Task" [ 782.741653] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.749790] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52154640-0ed6-7c4d-80f6-a0b0593fea13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.920068] env[61995]: DEBUG nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.941677] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.941677] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.941677] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.941677] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.941677] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.941677] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.941677] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.942232] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.942592] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.942913] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.946017] env[61995]: DEBUG nova.virt.hardware [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.946017] env[61995]: DEBUG nova.scheduler.client.report [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.951829] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e41ba11-dcb7-4eae-a9c0-0eacbc6578f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.967409] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2691e3ef-92d5-44ad-a2db-7db0d1df3230 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.031128] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794446, 'name': Rename_Task, 'duration_secs': 0.14665} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.031446] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 783.031710] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-008e10bc-dc01-44b9-9b8f-72cd198830f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.038449] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 783.038449] env[61995]: value = "task-794447" [ 783.038449] env[61995]: _type = "Task" [ 783.038449] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.046199] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.191502] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c43a1ad6-ef6e-491f-a485-9f0d19bdf5db tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.411s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.254026] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52154640-0ed6-7c4d-80f6-a0b0593fea13, 'name': SearchDatastore_Task, 'duration_secs': 0.051537} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.254431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.254746] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.255060] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.255278] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.255646] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.256246] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-11c5aa56-7699-42a0-a8a8-152cc25a0849 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.271571] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.271803] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 783.272662] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9613415-fba9-4180-98ee-4bd0eb915ea5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.282607] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 783.282607] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5210c84b-5a41-cd6e-6b70-87ad1c80edb9" [ 783.282607] env[61995]: _type = "Task" [ 783.282607] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.292342] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5210c84b-5a41-cd6e-6b70-87ad1c80edb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.462682] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.465770] env[61995]: DEBUG nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 783.466697] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.800s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.468505] env[61995]: INFO nova.compute.claims [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.549514] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794447, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.727156] env[61995]: DEBUG nova.network.neutron [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Successfully updated port: 9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.776497] env[61995]: DEBUG nova.compute.manager [req-e54313d7-5e29-41bb-ab4a-e4179156985b req-702646be-0d6b-437f-90fa-b161c07283ff service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Received event network-vif-plugged-9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 783.776566] env[61995]: DEBUG oslo_concurrency.lockutils [req-e54313d7-5e29-41bb-ab4a-e4179156985b req-702646be-0d6b-437f-90fa-b161c07283ff service nova] Acquiring lock "eef1cf32-e71e-4696-8170-1023e3a5518d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.776749] env[61995]: DEBUG oslo_concurrency.lockutils [req-e54313d7-5e29-41bb-ab4a-e4179156985b req-702646be-0d6b-437f-90fa-b161c07283ff service nova] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.776914] env[61995]: DEBUG oslo_concurrency.lockutils [req-e54313d7-5e29-41bb-ab4a-e4179156985b req-702646be-0d6b-437f-90fa-b161c07283ff service nova] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.777098] env[61995]: DEBUG nova.compute.manager [req-e54313d7-5e29-41bb-ab4a-e4179156985b req-702646be-0d6b-437f-90fa-b161c07283ff service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] No waiting events found dispatching network-vif-plugged-9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 783.777269] env[61995]: WARNING nova.compute.manager [req-e54313d7-5e29-41bb-ab4a-e4179156985b req-702646be-0d6b-437f-90fa-b161c07283ff service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Received unexpected event network-vif-plugged-9116dcbf-d617-485e-afde-c82e908d5c15 for instance with vm_state building and task_state spawning. [ 783.793822] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5210c84b-5a41-cd6e-6b70-87ad1c80edb9, 'name': SearchDatastore_Task, 'duration_secs': 0.030744} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.794624] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f97ef371-b35e-420e-8c12-2d8d4c4cb15d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.800903] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 783.800903] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]528e17c8-1ddf-3e3b-d66e-64d38a1a820f" [ 783.800903] env[61995]: _type = "Task" [ 783.800903] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.810345] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528e17c8-1ddf-3e3b-d66e-64d38a1a820f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.978109] env[61995]: DEBUG nova.compute.utils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.979671] env[61995]: DEBUG nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.979849] env[61995]: DEBUG nova.network.neutron [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 784.051762] env[61995]: DEBUG oslo_vmware.api [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794447, 'name': PowerOnVM_Task, 'duration_secs': 0.587009} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.054417] env[61995]: DEBUG nova.policy [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03c34fda698048d4ac4040bfa553606d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd4b615cdd3044a18c62dde884955839', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 784.056191] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 784.056448] env[61995]: INFO nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Took 8.17 seconds to spawn the instance on the hypervisor. [ 784.058947] env[61995]: DEBUG nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 784.058947] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8f8d56-c26b-4138-91dd-ee35eb90c3c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.138630] env[61995]: DEBUG nova.compute.manager [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Received event network-changed-9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.138863] env[61995]: DEBUG nova.compute.manager [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Refreshing instance network info cache due to event network-changed-9551a93f-ec41-48c7-b25a-aa038eca7341. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.139106] env[61995]: DEBUG oslo_concurrency.lockutils [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] Acquiring lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.139252] env[61995]: DEBUG oslo_concurrency.lockutils [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] Acquired lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.139418] env[61995]: DEBUG nova.network.neutron [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Refreshing network info cache for port 9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.229791] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.230045] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquired lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.230745] env[61995]: DEBUG nova.network.neutron [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.314411] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528e17c8-1ddf-3e3b-d66e-64d38a1a820f, 'name': SearchDatastore_Task, 'duration_secs': 0.018168} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.314689] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.314979] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] f9c1424f-6096-481c-aa91-acc25bfd8f6e/f9c1424f-6096-481c-aa91-acc25bfd8f6e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.315249] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-435f44db-d2ac-466a-9755-c322b64b593f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.322918] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 784.322918] env[61995]: value = "task-794448" [ 784.322918] env[61995]: _type = "Task" [ 784.322918] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.342620] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794448, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.488075] env[61995]: DEBUG nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 784.568722] env[61995]: DEBUG nova.network.neutron [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Successfully created port: 647ca67a-7528-4c2a-8500-6dac05d4ed89 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.582827] env[61995]: INFO nova.compute.manager [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Took 35.73 seconds to build instance. [ 784.817574] env[61995]: DEBUG nova.network.neutron [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.832169] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.832492] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.832709] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.832900] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.833094] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.844704] env[61995]: INFO nova.compute.manager [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Terminating instance [ 784.848161] env[61995]: DEBUG nova.compute.manager [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.848438] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 784.852745] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0bc920-0973-4737-b6df-50bc9b1a07fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.862370] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794448, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.867699] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 784.867962] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-034a83da-c7e7-42e3-81c3-a19870711497 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.875940] env[61995]: DEBUG oslo_vmware.api [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 784.875940] env[61995]: value = "task-794449" [ 784.875940] env[61995]: _type = "Task" [ 784.875940] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.888819] env[61995]: DEBUG oslo_vmware.api [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.955460] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7de4d3-d768-4ba7-a6c2-b1ad2e057863 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.964335] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b26a92-2c89-4b04-a27b-dcd440d8394d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.007464] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab1b544-016d-4f7b-9583-56cbf953ce2e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.017286] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd443fc-03df-4db3-a633-f858d54108ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.034630] env[61995]: DEBUG nova.compute.provider_tree [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.044949] env[61995]: DEBUG nova.network.neutron [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Updating instance_info_cache with network_info: [{"id": "9116dcbf-d617-485e-afde-c82e908d5c15", "address": "fa:16:3e:bc:1a:0f", "network": {"id": "bef08ad8-f37d-4b6a-a91c-e475beb866e7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1845121281-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "079a11ce311d40f9b8344fcce2e03683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9116dcbf-d6", "ovs_interfaceid": "9116dcbf-d617-485e-afde-c82e908d5c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.085515] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bcf182c8-3b51-4cfc-ab43-ab11ebbaf1f7 tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.786s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.122277] env[61995]: DEBUG nova.network.neutron [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updated VIF entry in instance network info cache for port 9551a93f-ec41-48c7-b25a-aa038eca7341. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.122679] env[61995]: DEBUG nova.network.neutron [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updating instance_info_cache with network_info: [{"id": "9551a93f-ec41-48c7-b25a-aa038eca7341", "address": "fa:16:3e:d2:6c:28", "network": {"id": "616dc7e4-858b-4038-bfd9-30dc98a41f73", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-733696429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb4715780da643999551e223caa16065", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551a93f-ec", "ovs_interfaceid": "9551a93f-ec41-48c7-b25a-aa038eca7341", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.339540] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794448, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569812} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.339856] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] f9c1424f-6096-481c-aa91-acc25bfd8f6e/f9c1424f-6096-481c-aa91-acc25bfd8f6e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 785.340962] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.340962] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cedfde6a-1a65-44b4-a35d-36cfd8306f9a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.347272] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 785.347272] env[61995]: value = "task-794450" [ 785.347272] env[61995]: _type = "Task" [ 785.347272] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.355818] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794450, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.390870] env[61995]: DEBUG oslo_vmware.api [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794449, 'name': PowerOffVM_Task, 'duration_secs': 0.234013} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.391119] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 785.391344] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 785.391704] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11330728-5954-4e6d-abe8-805161a5cde2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.469701] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 785.469701] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 785.469701] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Deleting the datastore file [datastore2] c51184b6-d2c7-49ff-b061-ab0f9a69cdbf {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.469701] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40fe44ce-277e-45e8-a662-dc9ec9092817 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.477799] env[61995]: DEBUG oslo_vmware.api [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for the task: (returnval){ [ 785.477799] env[61995]: value = "task-794452" [ 785.477799] env[61995]: _type = "Task" [ 785.477799] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.488399] env[61995]: DEBUG oslo_vmware.api [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.514102] env[61995]: DEBUG nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 785.537393] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.537718] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.537921] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.538275] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.538488] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.538786] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.539099] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.539340] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.539589] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.539810] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.540084] env[61995]: DEBUG nova.virt.hardware [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.541136] env[61995]: DEBUG nova.scheduler.client.report [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.545763] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a165aad1-0b78-422a-a1e5-4e040076062f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.549756] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Releasing lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.550136] env[61995]: DEBUG nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Instance network_info: |[{"id": "9116dcbf-d617-485e-afde-c82e908d5c15", "address": "fa:16:3e:bc:1a:0f", "network": {"id": "bef08ad8-f37d-4b6a-a91c-e475beb866e7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1845121281-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "079a11ce311d40f9b8344fcce2e03683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9116dcbf-d6", "ovs_interfaceid": "9116dcbf-d617-485e-afde-c82e908d5c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 785.550595] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:1a:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604c9724-b4ef-4393-a76e-eb4a2b510796', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9116dcbf-d617-485e-afde-c82e908d5c15', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.559331] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Creating folder: Project (079a11ce311d40f9b8344fcce2e03683). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.560867] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7699cbc-fbd7-4868-89ee-e80ba967190d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.567602] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe122cfc-462c-45b6-b7e2-ac96aa0ed50f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.575934] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Created folder: Project (079a11ce311d40f9b8344fcce2e03683) in parent group-v185203. [ 785.576238] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Creating folder: Instances. Parent ref: group-v185282. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.584780] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d8b03c4-187a-4cad-a140-495765c62bde {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.588775] env[61995]: DEBUG nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.596765] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Created folder: Instances in parent group-v185282. [ 785.597093] env[61995]: DEBUG oslo.service.loopingcall [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.597248] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 785.597477] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ddcd600e-31ba-4c18-be94-c59139a88dfa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.620550] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.620550] env[61995]: value = "task-794455" [ 785.620550] env[61995]: _type = "Task" [ 785.620550] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.625655] env[61995]: DEBUG oslo_concurrency.lockutils [req-15b0eeb3-0f7a-4d31-a666-8aa2f49d997b req-36248ea2-23ea-477c-92ab-c5e4b1929226 service nova] Releasing lock "refresh_cache-c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.629638] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794455, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.681152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.685022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.685022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.685022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.685022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.685022] env[61995]: INFO nova.compute.manager [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Terminating instance [ 785.693295] env[61995]: DEBUG nova.compute.manager [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 785.693652] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 785.694515] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b4cf176-2035-40e4-a06b-e64991a77c6b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.705206] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 785.706295] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5906bcda-1653-474d-8e71-154b9bb2fed6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.714966] env[61995]: DEBUG oslo_vmware.api [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 785.714966] env[61995]: value = "task-794456" [ 785.714966] env[61995]: _type = "Task" [ 785.714966] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.726125] env[61995]: DEBUG oslo_vmware.api [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794456, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.868928] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794450, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086016} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.870406] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.871907] env[61995]: DEBUG nova.compute.manager [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Received event network-changed-9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.872087] env[61995]: DEBUG nova.compute.manager [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Refreshing instance network info cache due to event network-changed-9116dcbf-d617-485e-afde-c82e908d5c15. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 785.872333] env[61995]: DEBUG oslo_concurrency.lockutils [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] Acquiring lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.872516] env[61995]: DEBUG oslo_concurrency.lockutils [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] Acquired lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.872675] env[61995]: DEBUG nova.network.neutron [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Refreshing network info cache for port 9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 785.874693] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a692edda-3c38-4398-84c0-fd294de53cb7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.908302] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] f9c1424f-6096-481c-aa91-acc25bfd8f6e/f9c1424f-6096-481c-aa91-acc25bfd8f6e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.908536] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8a6f3eb-3bfc-4571-b283-3234bb46a604 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.932597] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 785.932597] env[61995]: value = "task-794457" [ 785.932597] env[61995]: _type = "Task" [ 785.932597] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.946714] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794457, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.990034] env[61995]: DEBUG oslo_vmware.api [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Task: {'id': task-794452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234005} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.990394] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.990609] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 785.990796] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 785.991047] env[61995]: INFO nova.compute.manager [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 785.991340] env[61995]: DEBUG oslo.service.loopingcall [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.991563] env[61995]: DEBUG nova.compute.manager [-] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.991671] env[61995]: DEBUG nova.network.neutron [-] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 786.050834] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.051526] env[61995]: DEBUG nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.056635] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.258s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.058819] env[61995]: INFO nova.compute.claims [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.117844] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.132231] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794455, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.225482] env[61995]: DEBUG oslo_vmware.api [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794456, 'name': PowerOffVM_Task, 'duration_secs': 0.255418} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.225845] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 786.226057] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 786.226343] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f1855aa-2da4-4613-9a53-20019736d8e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.318543] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 786.319534] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 786.319534] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Deleting the datastore file [datastore2] f6db353f-f7b5-4ba4-8f70-f3b65281da93 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.319534] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-373b1953-bc0a-4957-b20d-1f5a6fbd1084 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.327947] env[61995]: DEBUG oslo_vmware.api [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for the task: (returnval){ [ 786.327947] env[61995]: value = "task-794459" [ 786.327947] env[61995]: _type = "Task" [ 786.327947] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.337237] env[61995]: DEBUG oslo_vmware.api [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794459, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.443900] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794457, 'name': ReconfigVM_Task, 'duration_secs': 0.460455} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.443900] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Reconfigured VM instance instance-0000002e to attach disk [datastore2] f9c1424f-6096-481c-aa91-acc25bfd8f6e/f9c1424f-6096-481c-aa91-acc25bfd8f6e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.444354] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea10f9c3-463d-4302-aa27-3766f369df53 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.452245] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 786.452245] env[61995]: value = "task-794460" [ 786.452245] env[61995]: _type = "Task" [ 786.452245] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.461507] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794460, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.563910] env[61995]: DEBUG nova.compute.utils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.570024] env[61995]: DEBUG nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.570024] env[61995]: DEBUG nova.network.neutron [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 786.634123] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794455, 'name': CreateVM_Task, 'duration_secs': 0.526107} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.634457] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.635880] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.636510] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.637228] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.638131] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-238218c3-0e2a-4919-8ce9-4d5ec57215d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.644872] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 786.644872] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5215ecf3-f09c-aa10-32f4-3c1986a10900" [ 786.644872] env[61995]: _type = "Task" [ 786.644872] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.655362] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5215ecf3-f09c-aa10-32f4-3c1986a10900, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.693649] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a2b5-9368-268b-7c7d-bc956f781b08/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 786.698018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fe90e9-f415-40cf-b232-835961cc350b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.708260] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a2b5-9368-268b-7c7d-bc956f781b08/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 786.709980] env[61995]: ERROR oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a2b5-9368-268b-7c7d-bc956f781b08/disk-0.vmdk due to incomplete transfer. [ 786.709980] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-93f39c91-f73c-4751-89b0-f889be80c4d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.722016] env[61995]: DEBUG oslo_vmware.rw_handles [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a2b5-9368-268b-7c7d-bc956f781b08/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 786.722016] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Uploaded image c24b00a5-c673-4ee2-8bcb-389e5191f7fe to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 786.724090] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 786.726125] env[61995]: DEBUG nova.policy [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fda120058c42481f9b06de4b03539f44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df7381568d0e44fbb8bb161e9d5e8bb7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 786.728388] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0f9444fd-7d93-4de4-865f-b23d2c10d312 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.737361] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 786.737361] env[61995]: value = "task-794461" [ 786.737361] env[61995]: _type = "Task" [ 786.737361] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.752688] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794461, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.758028] env[61995]: DEBUG nova.network.neutron [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Updated VIF entry in instance network info cache for port 9116dcbf-d617-485e-afde-c82e908d5c15. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 786.758028] env[61995]: DEBUG nova.network.neutron [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Updating instance_info_cache with network_info: [{"id": "9116dcbf-d617-485e-afde-c82e908d5c15", "address": "fa:16:3e:bc:1a:0f", "network": {"id": "bef08ad8-f37d-4b6a-a91c-e475beb866e7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1845121281-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "079a11ce311d40f9b8344fcce2e03683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9116dcbf-d6", "ovs_interfaceid": "9116dcbf-d617-485e-afde-c82e908d5c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.840180] env[61995]: DEBUG oslo_vmware.api [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Task: {'id': task-794459, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247528} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.840637] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.841078] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 786.841409] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 786.842055] env[61995]: INFO nova.compute.manager [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Took 1.15 seconds to destroy the instance on the hypervisor. [ 786.842505] env[61995]: DEBUG oslo.service.loopingcall [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.842820] env[61995]: DEBUG nova.compute.manager [-] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.843042] env[61995]: DEBUG nova.network.neutron [-] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 786.965432] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794460, 'name': Rename_Task, 'duration_secs': 0.206293} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.969194] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.969194] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d10de3d3-e92e-4763-af11-2a88b9cfdde9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.975096] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 786.975096] env[61995]: value = "task-794462" [ 786.975096] env[61995]: _type = "Task" [ 786.975096] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.991406] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.077080] env[61995]: DEBUG nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.164052] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5215ecf3-f09c-aa10-32f4-3c1986a10900, 'name': SearchDatastore_Task, 'duration_secs': 0.011153} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.166114] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.166224] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.166472] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.166657] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.166871] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.167379] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f0fa72d-6e67-4e22-9400-8d68486b57bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.184480] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.184480] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 787.184923] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec51895-2a52-4afd-966f-486b39fa9e4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.192410] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 787.192410] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ff7766-06f3-e73f-e9fd-028b77356fad" [ 787.192410] env[61995]: _type = "Task" [ 787.192410] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.201787] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ff7766-06f3-e73f-e9fd-028b77356fad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.249821] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794461, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.259875] env[61995]: DEBUG oslo_concurrency.lockutils [req-ed0e605b-db09-41d6-8a36-b102315fd8c7 req-7d17e0e0-d212-4fb0-83ec-ce4ccff7eaac service nova] Releasing lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.270026] env[61995]: DEBUG nova.compute.manager [req-24a7e170-b8f6-4b95-bb25-5f43363170f5 req-6acdec5a-7dfb-476b-b3ae-fc464d417cab service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Received event network-vif-deleted-b2df31cb-ec39-4c6d-8011-9ab55731a25f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.270245] env[61995]: INFO nova.compute.manager [req-24a7e170-b8f6-4b95-bb25-5f43363170f5 req-6acdec5a-7dfb-476b-b3ae-fc464d417cab service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Neutron deleted interface b2df31cb-ec39-4c6d-8011-9ab55731a25f; detaching it from the instance and deleting it from the info cache [ 787.270546] env[61995]: DEBUG nova.network.neutron [req-24a7e170-b8f6-4b95-bb25-5f43363170f5 req-6acdec5a-7dfb-476b-b3ae-fc464d417cab service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.272721] env[61995]: DEBUG nova.network.neutron [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Successfully updated port: 647ca67a-7528-4c2a-8500-6dac05d4ed89 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.496157] env[61995]: DEBUG nova.network.neutron [-] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.504039] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794462, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.515201] env[61995]: DEBUG nova.network.neutron [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Successfully created port: 2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.520431] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1604a938-d1b5-423a-9a5a-758c6a8d6635 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.529337] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a009b8-cec2-473b-b801-357c4838415b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.569154] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38375caf-c321-487f-89d2-bec268398a72 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.578358] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6489b853-7588-427c-ae80-8d5ff5044ab9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.599443] env[61995]: DEBUG nova.compute.provider_tree [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.667809] env[61995]: DEBUG nova.network.neutron [-] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.703655] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ff7766-06f3-e73f-e9fd-028b77356fad, 'name': SearchDatastore_Task, 'duration_secs': 0.010049} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.704572] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ffe63ba-ed61-4d6e-ac53-ebaf2d23e8e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.712365] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 787.712365] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52be4795-7a66-5647-aa25-f7a225d75d60" [ 787.712365] env[61995]: _type = "Task" [ 787.712365] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.725624] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52be4795-7a66-5647-aa25-f7a225d75d60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.750065] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794461, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.775906] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "refresh_cache-d056dadd-492d-4a4d-abc2-fc5f113f865b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.776109] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquired lock "refresh_cache-d056dadd-492d-4a4d-abc2-fc5f113f865b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.776388] env[61995]: DEBUG nova.network.neutron [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 787.778961] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b609e604-ecf7-4dc5-92f9-9a721b0b541e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.791500] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad59a5e5-da71-401d-a365-45794b8a7a2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.829145] env[61995]: DEBUG nova.compute.manager [req-24a7e170-b8f6-4b95-bb25-5f43363170f5 req-6acdec5a-7dfb-476b-b3ae-fc464d417cab service nova] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Detach interface failed, port_id=b2df31cb-ec39-4c6d-8011-9ab55731a25f, reason: Instance f6db353f-f7b5-4ba4-8f70-f3b65281da93 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 787.906041] env[61995]: DEBUG nova.compute.manager [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Received event network-vif-plugged-647ca67a-7528-4c2a-8500-6dac05d4ed89 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.906041] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] Acquiring lock "d056dadd-492d-4a4d-abc2-fc5f113f865b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.906041] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.906270] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.906497] env[61995]: DEBUG nova.compute.manager [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] No waiting events found dispatching network-vif-plugged-647ca67a-7528-4c2a-8500-6dac05d4ed89 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 787.907310] env[61995]: WARNING nova.compute.manager [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Received unexpected event network-vif-plugged-647ca67a-7528-4c2a-8500-6dac05d4ed89 for instance with vm_state building and task_state spawning. [ 787.907310] env[61995]: DEBUG nova.compute.manager [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Received event network-vif-deleted-9551a93f-ec41-48c7-b25a-aa038eca7341 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.907310] env[61995]: DEBUG nova.compute.manager [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Received event network-changed-647ca67a-7528-4c2a-8500-6dac05d4ed89 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.907310] env[61995]: DEBUG nova.compute.manager [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Refreshing instance network info cache due to event network-changed-647ca67a-7528-4c2a-8500-6dac05d4ed89. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 787.907607] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] Acquiring lock "refresh_cache-d056dadd-492d-4a4d-abc2-fc5f113f865b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.991704] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794462, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.006252] env[61995]: INFO nova.compute.manager [-] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Took 2.01 seconds to deallocate network for instance. [ 788.105518] env[61995]: DEBUG nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.107118] env[61995]: DEBUG nova.scheduler.client.report [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.133972] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.134136] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.134354] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.134603] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.134801] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.134988] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.135338] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.135578] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.136540] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.136879] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.136988] env[61995]: DEBUG nova.virt.hardware [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.138411] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5682f35-0f03-48c3-92ad-e344a37eeba9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.148751] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01dcc259-9482-4516-80cd-9db708a72592 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.171743] env[61995]: INFO nova.compute.manager [-] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Took 1.33 seconds to deallocate network for instance. [ 788.225744] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52be4795-7a66-5647-aa25-f7a225d75d60, 'name': SearchDatastore_Task, 'duration_secs': 0.023986} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.225800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.226128] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] eef1cf32-e71e-4696-8170-1023e3a5518d/eef1cf32-e71e-4696-8170-1023e3a5518d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.226797] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a518179d-b48b-4e52-8970-582893a2a1f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.236031] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 788.236031] env[61995]: value = "task-794463" [ 788.236031] env[61995]: _type = "Task" [ 788.236031] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.247926] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794463, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.254017] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794461, 'name': Destroy_Task, 'duration_secs': 1.429372} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.254429] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Destroyed the VM [ 788.254793] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 788.255070] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ac254e7c-f5b0-4891-a4d7-c87b2fc7563a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.263529] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 788.263529] env[61995]: value = "task-794464" [ 788.263529] env[61995]: _type = "Task" [ 788.263529] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.273061] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794464, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.339394] env[61995]: DEBUG nova.network.neutron [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.494150] env[61995]: DEBUG oslo_vmware.api [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794462, 'name': PowerOnVM_Task, 'duration_secs': 1.202573} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.494150] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 788.494150] env[61995]: INFO nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Took 10.05 seconds to spawn the instance on the hypervisor. [ 788.494150] env[61995]: DEBUG nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.494916] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3791b4f0-e9c3-4199-9410-c1225f82f3d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.517489] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.611629] env[61995]: DEBUG nova.network.neutron [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Updating instance_info_cache with network_info: [{"id": "647ca67a-7528-4c2a-8500-6dac05d4ed89", "address": "fa:16:3e:6e:9e:2e", "network": {"id": "4aba1a8a-fe92-47e4-a70a-0506f75c7521", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-38656509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd4b615cdd3044a18c62dde884955839", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap647ca67a-75", "ovs_interfaceid": "647ca67a-7528-4c2a-8500-6dac05d4ed89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.613951] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.615215] env[61995]: DEBUG nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.619112] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.790s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.619428] env[61995]: DEBUG nova.objects.instance [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lazy-loading 'resources' on Instance uuid 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.678738] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.750553] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794463, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509235} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.750844] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] eef1cf32-e71e-4696-8170-1023e3a5518d/eef1cf32-e71e-4696-8170-1023e3a5518d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 788.751083] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.751341] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fd9d1c6-c86e-4c9a-a9b2-5d4de86e1d5a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.759495] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 788.759495] env[61995]: value = "task-794465" [ 788.759495] env[61995]: _type = "Task" [ 788.759495] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.771898] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.777442] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794464, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.017401] env[61995]: INFO nova.compute.manager [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Took 38.54 seconds to build instance. [ 789.120280] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Releasing lock "refresh_cache-d056dadd-492d-4a4d-abc2-fc5f113f865b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.121080] env[61995]: DEBUG nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Instance network_info: |[{"id": "647ca67a-7528-4c2a-8500-6dac05d4ed89", "address": "fa:16:3e:6e:9e:2e", "network": {"id": "4aba1a8a-fe92-47e4-a70a-0506f75c7521", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-38656509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd4b615cdd3044a18c62dde884955839", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap647ca67a-75", "ovs_interfaceid": "647ca67a-7528-4c2a-8500-6dac05d4ed89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.121080] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] Acquired lock "refresh_cache-d056dadd-492d-4a4d-abc2-fc5f113f865b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.121420] env[61995]: DEBUG nova.network.neutron [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Refreshing network info cache for port 647ca67a-7528-4c2a-8500-6dac05d4ed89 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.123445] env[61995]: DEBUG nova.compute.utils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.124905] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:9e:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '41f66e20-fd86-4158-bbdc-7a150e85e844', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '647ca67a-7528-4c2a-8500-6dac05d4ed89', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.132778] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Creating folder: Project (dd4b615cdd3044a18c62dde884955839). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.136502] env[61995]: DEBUG nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.136707] env[61995]: DEBUG nova.network.neutron [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 789.138396] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a92e58a-9d1a-475e-bbaf-bb91d5de5a00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.152844] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Created folder: Project (dd4b615cdd3044a18c62dde884955839) in parent group-v185203. [ 789.153085] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Creating folder: Instances. Parent ref: group-v185285. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.153330] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-904ab0b1-3cc6-457c-bb56-d3e256666dbe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.166694] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Created folder: Instances in parent group-v185285. [ 789.166935] env[61995]: DEBUG oslo.service.loopingcall [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.167149] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 789.167471] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8992e050-03e9-45af-8106-88657387c728 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.191980] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.191980] env[61995]: value = "task-794468" [ 789.191980] env[61995]: _type = "Task" [ 789.191980] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.200272] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794468, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.243805] env[61995]: DEBUG nova.policy [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16bf12381242949167ffc13d3aa404', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e0840f3a4d24ea78c35c4b273f1e413', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 789.274372] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066163} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.275190] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.275856] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea25253-2521-4a7e-8a0c-9bdaaecfdf7e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.281540] env[61995]: DEBUG oslo_vmware.api [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794464, 'name': RemoveSnapshot_Task, 'duration_secs': 0.776547} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.282265] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 789.282562] env[61995]: INFO nova.compute.manager [None req-cd46e14b-0cd1-49ff-9b3a-30563ef4c0ec tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Took 18.71 seconds to snapshot the instance on the hypervisor. [ 789.303571] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] eef1cf32-e71e-4696-8170-1023e3a5518d/eef1cf32-e71e-4696-8170-1023e3a5518d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.306778] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6ad0540-0c13-4f5f-8b2c-36a90d45432f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.328928] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 789.328928] env[61995]: value = "task-794469" [ 789.328928] env[61995]: _type = "Task" [ 789.328928] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.344945] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794469, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.521655] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f659df7d-eacf-43da-9d7d-c5a9e1d44e2d tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.406s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.632898] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1df5cb6-09f3-4923-8956-3fbfeb1a707c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.638292] env[61995]: DEBUG nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.645501] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8a7ff9-a597-4134-af2c-f70ef74db725 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.678192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88eccfd3-8263-4da6-a265-7eedbd87c546 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.689488] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7888d35-b7ff-4203-b7d9-dbc6c222a08c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.708580] env[61995]: DEBUG nova.compute.provider_tree [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.713092] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794468, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.840034] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794469, 'name': ReconfigVM_Task, 'duration_secs': 0.298956} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.840798] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Reconfigured VM instance instance-0000002f to attach disk [datastore1] eef1cf32-e71e-4696-8170-1023e3a5518d/eef1cf32-e71e-4696-8170-1023e3a5518d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.841687] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77eb8214-49e3-453a-968d-e033f9f733dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.848784] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 789.848784] env[61995]: value = "task-794470" [ 789.848784] env[61995]: _type = "Task" [ 789.848784] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.859933] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794470, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.024659] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.085397] env[61995]: DEBUG nova.network.neutron [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Successfully updated port: 2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.093141] env[61995]: DEBUG nova.network.neutron [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Successfully created port: 6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.200342] env[61995]: DEBUG nova.network.neutron [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Updated VIF entry in instance network info cache for port 647ca67a-7528-4c2a-8500-6dac05d4ed89. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.200342] env[61995]: DEBUG nova.network.neutron [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Updating instance_info_cache with network_info: [{"id": "647ca67a-7528-4c2a-8500-6dac05d4ed89", "address": "fa:16:3e:6e:9e:2e", "network": {"id": "4aba1a8a-fe92-47e4-a70a-0506f75c7521", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-38656509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd4b615cdd3044a18c62dde884955839", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap647ca67a-75", "ovs_interfaceid": "647ca67a-7528-4c2a-8500-6dac05d4ed89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.207085] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794468, 'name': CreateVM_Task, 'duration_secs': 0.848515} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.207252] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 790.207924] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.208142] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.208480] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.208736] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae68ec5c-841e-498e-aca8-f48c8b628465 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.214598] env[61995]: DEBUG nova.scheduler.client.report [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.218095] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 790.218095] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5274cee1-ee35-fb67-0880-4bf1c23b5bf2" [ 790.218095] env[61995]: _type = "Task" [ 790.218095] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.227848] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5274cee1-ee35-fb67-0880-4bf1c23b5bf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.268084] env[61995]: DEBUG nova.compute.manager [req-dd09c60b-cea6-48c4-9471-51531c6baf4e req-3d02daa5-66ed-47f9-b999-ccc3416e2cde service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Received event network-vif-plugged-2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.268337] env[61995]: DEBUG oslo_concurrency.lockutils [req-dd09c60b-cea6-48c4-9471-51531c6baf4e req-3d02daa5-66ed-47f9-b999-ccc3416e2cde service nova] Acquiring lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.268553] env[61995]: DEBUG oslo_concurrency.lockutils [req-dd09c60b-cea6-48c4-9471-51531c6baf4e req-3d02daa5-66ed-47f9-b999-ccc3416e2cde service nova] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.268726] env[61995]: DEBUG oslo_concurrency.lockutils [req-dd09c60b-cea6-48c4-9471-51531c6baf4e req-3d02daa5-66ed-47f9-b999-ccc3416e2cde service nova] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.268899] env[61995]: DEBUG nova.compute.manager [req-dd09c60b-cea6-48c4-9471-51531c6baf4e req-3d02daa5-66ed-47f9-b999-ccc3416e2cde service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] No waiting events found dispatching network-vif-plugged-2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.269215] env[61995]: WARNING nova.compute.manager [req-dd09c60b-cea6-48c4-9471-51531c6baf4e req-3d02daa5-66ed-47f9-b999-ccc3416e2cde service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Received unexpected event network-vif-plugged-2c08d25b-1e7a-4f2d-8715-0e64edc11b01 for instance with vm_state building and task_state spawning. [ 790.359366] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794470, 'name': Rename_Task, 'duration_secs': 0.13979} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.360016] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 790.360016] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee3d6be4-3bf0-420a-ad9e-a81849bf8f38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.366904] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 790.366904] env[61995]: value = "task-794471" [ 790.366904] env[61995]: _type = "Task" [ 790.366904] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.375418] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.556995] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.590760] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.590915] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.591092] env[61995]: DEBUG nova.network.neutron [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 790.650806] env[61995]: DEBUG nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.677540] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.677917] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.678232] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.678674] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.678856] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.679035] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.679261] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.679428] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.679605] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.679772] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.679952] env[61995]: DEBUG nova.virt.hardware [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.680825] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc973c7e-85db-4931-bf95-26b95a42c6a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.689396] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5620c227-2163-4245-b0f0-7cc77945bb3d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.703936] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d11159b-b75a-4855-9c53-43e8546d5ac1 req-a502d9e9-cd6e-426c-85c7-34696f0d4ac6 service nova] Releasing lock "refresh_cache-d056dadd-492d-4a4d-abc2-fc5f113f865b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.720457] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.100s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.722682] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.643s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.723852] env[61995]: INFO nova.compute.claims [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.740988] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5274cee1-ee35-fb67-0880-4bf1c23b5bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.02074} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.740988] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.741289] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.741367] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.741559] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.742420] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.742420] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0bdbf9e-f2f6-44f8-aded-edafb6585da3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.750760] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.751582] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 790.751713] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c7b86f8-87ba-4c96-8bb6-f750d9db0101 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.755074] env[61995]: INFO nova.scheduler.client.report [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Deleted allocations for instance 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa [ 790.762916] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 790.762916] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c8383b-2872-cc0a-d294-5338af3ca5dc" [ 790.762916] env[61995]: _type = "Task" [ 790.762916] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.772039] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c8383b-2872-cc0a-d294-5338af3ca5dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.878812] env[61995]: DEBUG oslo_vmware.api [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794471, 'name': PowerOnVM_Task, 'duration_secs': 0.440356} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.879211] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 790.879551] env[61995]: INFO nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Took 7.96 seconds to spawn the instance on the hypervisor. [ 790.879809] env[61995]: DEBUG nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.880739] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785959f6-8d50-46b3-872a-4883ba569a66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.904399] env[61995]: DEBUG nova.compute.manager [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Received event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.904606] env[61995]: DEBUG nova.compute.manager [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing instance network info cache due to event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 790.904849] env[61995]: DEBUG oslo_concurrency.lockutils [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] Acquiring lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.905010] env[61995]: DEBUG oslo_concurrency.lockutils [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] Acquired lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.905355] env[61995]: DEBUG nova.network.neutron [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 791.130138] env[61995]: DEBUG nova.network.neutron [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.271354] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb0dc4bd-0a8e-42f6-bc9c-59735e6bace5 tempest-ServerAddressesTestJSON-1271430944 tempest-ServerAddressesTestJSON-1271430944-project-member] Lock "0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.406s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.285732] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c8383b-2872-cc0a-d294-5338af3ca5dc, 'name': SearchDatastore_Task, 'duration_secs': 0.01143} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.286903] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ec15727-778b-42b6-aa52-51d60b3c23f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.296021] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 791.296021] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52793e60-93f6-bace-295c-60c956cded3b" [ 791.296021] env[61995]: _type = "Task" [ 791.296021] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.304448] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52793e60-93f6-bace-295c-60c956cded3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.351611] env[61995]: DEBUG nova.network.neutron [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Updating instance_info_cache with network_info: [{"id": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "address": "fa:16:3e:c1:25:58", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c08d25b-1e", "ovs_interfaceid": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.405170] env[61995]: INFO nova.compute.manager [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Took 35.07 seconds to build instance. [ 791.657265] env[61995]: DEBUG nova.network.neutron [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Successfully updated port: 6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.810291] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52793e60-93f6-bace-295c-60c956cded3b, 'name': SearchDatastore_Task, 'duration_secs': 0.025966} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.810580] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.810839] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] d056dadd-492d-4a4d-abc2-fc5f113f865b/d056dadd-492d-4a4d-abc2-fc5f113f865b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 791.811134] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d68925c0-8052-4f64-8625-95ffa5742f2d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.820567] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 791.820567] env[61995]: value = "task-794472" [ 791.820567] env[61995]: _type = "Task" [ 791.820567] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.832096] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794472, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.855703] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Releasing lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.855703] env[61995]: DEBUG nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Instance network_info: |[{"id": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "address": "fa:16:3e:c1:25:58", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c08d25b-1e", "ovs_interfaceid": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 791.855703] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:25:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '352165bb-004f-4180-9627-3a275dbe18af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c08d25b-1e7a-4f2d-8715-0e64edc11b01', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 791.864902] env[61995]: DEBUG oslo.service.loopingcall [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.865233] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 791.869180] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9439083-3c95-4fae-98a0-60025386d731 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.888877] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 791.888877] env[61995]: value = "task-794473" [ 791.888877] env[61995]: _type = "Task" [ 791.888877] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.900109] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794473, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.900109] env[61995]: DEBUG nova.network.neutron [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updated VIF entry in instance network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 791.900109] env[61995]: DEBUG nova.network.neutron [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.911128] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4878f407-a748-491f-990f-999b45e7bed6 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.858s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.164026] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.164176] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.164331] env[61995]: DEBUG nova.network.neutron [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 792.258827] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8595c975-2ab4-497e-a0e3-54b519b71ede {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.266270] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2acd94a-1f23-41e5-be2b-7f04ff46ad40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.301462] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93000512-dc8e-4e50-a73e-c765d24018c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.310755] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017f2444-7029-49ee-b2d1-68982945c518 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.327812] env[61995]: DEBUG nova.compute.provider_tree [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.338876] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794472, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.402745] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794473, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.405452] env[61995]: DEBUG oslo_concurrency.lockutils [req-83aab3be-949f-4514-bd03-76379d2b5901 req-d9f2ae98-13a7-4faa-b655-20e8d42d3dcc service nova] Releasing lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.413624] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.587169] env[61995]: DEBUG nova.compute.manager [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Received event network-changed-2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.587435] env[61995]: DEBUG nova.compute.manager [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Refreshing instance network info cache due to event network-changed-2c08d25b-1e7a-4f2d-8715-0e64edc11b01. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 792.587613] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Acquiring lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.587759] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Acquired lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.588033] env[61995]: DEBUG nova.network.neutron [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Refreshing network info cache for port 2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.599775] env[61995]: DEBUG nova.compute.manager [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.602535] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28842712-b9fc-4a41-a0e2-2542d075d869 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.699905] env[61995]: DEBUG nova.network.neutron [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.834512] env[61995]: DEBUG nova.scheduler.client.report [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.841540] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794472, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.584827} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.841745] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] d056dadd-492d-4a4d-abc2-fc5f113f865b/d056dadd-492d-4a4d-abc2-fc5f113f865b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 792.841965] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.842229] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed05c943-e4e3-4658-9526-44e036817a38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.849609] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 792.849609] env[61995]: value = "task-794474" [ 792.849609] env[61995]: _type = "Task" [ 792.849609] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.859593] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.905529] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794473, 'name': CreateVM_Task, 'duration_secs': 0.6417} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.905706] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 792.906451] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.906623] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.906963] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 792.907236] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5741516-27ce-4abe-8544-9ff7c9c4e68d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.910069] env[61995]: DEBUG nova.network.neutron [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee71b78-02", "ovs_interfaceid": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.914283] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 792.914283] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ebdfa4-eaef-c162-8336-af4f60888388" [ 792.914283] env[61995]: _type = "Task" [ 792.914283] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.926823] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ebdfa4-eaef-c162-8336-af4f60888388, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.939869] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.120645] env[61995]: INFO nova.compute.manager [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] instance snapshotting [ 793.124731] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439a1128-5653-4808-896f-7d8cb653e28f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.149583] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ea94ba-f6ad-4ee9-9356-636cce321378 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.173600] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "451e226c-5fb7-4042-a706-3e58b0606546" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.173861] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "451e226c-5fb7-4042-a706-3e58b0606546" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.342929] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.345645] env[61995]: DEBUG nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.350336] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 26.444s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.366987] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06678} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.367548] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.368763] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb1e363-93e3-4af9-a782-a91442a1dc60 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.397969] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] d056dadd-492d-4a4d-abc2-fc5f113f865b/d056dadd-492d-4a4d-abc2-fc5f113f865b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.399134] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d6973b2-e571-4c85-9e29-c293207454e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.417040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.417417] env[61995]: DEBUG nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Instance network_info: |[{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee71b78-02", "ovs_interfaceid": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 793.417964] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:b9:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff1f3320-df8e-49df-a412-9797a23bd173', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ee71b78-02c5-4f63-a764-d5f1bb61107c', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.426034] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating folder: Project (3e0840f3a4d24ea78c35c4b273f1e413). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 793.431506] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78ef1a4d-86bd-42c9-8b34-b19104f89f99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.436205] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 793.436205] env[61995]: value = "task-794475" [ 793.436205] env[61995]: _type = "Task" [ 793.436205] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.442789] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ebdfa4-eaef-c162-8336-af4f60888388, 'name': SearchDatastore_Task, 'duration_secs': 0.010389} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.449873] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.453015] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.453015] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.453015] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.453015] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.453015] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e42a231c-dd7c-4327-bb87-3f98cd92f994 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.455493] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Created folder: Project (3e0840f3a4d24ea78c35c4b273f1e413) in parent group-v185203. [ 793.455822] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating folder: Instances. Parent ref: group-v185289. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 793.456525] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f92e969b-a59b-4048-beb3-a3c546b4a1f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.462575] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.463917] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.464125] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 793.465872] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-600a2aea-2923-4851-b6e8-3dc6f62c7890 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.470021] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Created folder: Instances in parent group-v185289. [ 793.470021] env[61995]: DEBUG oslo.service.loopingcall [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.470021] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 793.470021] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0efdda3-80cc-4073-81fe-d41e99f1fc1d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.485976] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 793.485976] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5228edd6-f9eb-b812-2a92-0830919da01a" [ 793.485976] env[61995]: _type = "Task" [ 793.485976] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.490973] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.490973] env[61995]: value = "task-794478" [ 793.490973] env[61995]: _type = "Task" [ 793.490973] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.497459] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5228edd6-f9eb-b812-2a92-0830919da01a, 'name': SearchDatastore_Task, 'duration_secs': 0.012767} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.498642] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e7332a3-450b-4515-8b3b-8e99d024a7fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.504895] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794478, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.509983] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 793.509983] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]528bb40b-d09b-d0f2-f9e5-d9a414e7563f" [ 793.509983] env[61995]: _type = "Task" [ 793.509983] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.514451] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528bb40b-d09b-d0f2-f9e5-d9a414e7563f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.515311] env[61995]: DEBUG nova.network.neutron [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Updated VIF entry in instance network info cache for port 2c08d25b-1e7a-4f2d-8715-0e64edc11b01. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.516685] env[61995]: DEBUG nova.network.neutron [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Updating instance_info_cache with network_info: [{"id": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "address": "fa:16:3e:c1:25:58", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c08d25b-1e", "ovs_interfaceid": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.661216] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 793.661555] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0ad46780-1e07-413a-af26-fd70b205245e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.668836] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 793.668836] env[61995]: value = "task-794479" [ 793.668836] env[61995]: _type = "Task" [ 793.668836] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.680298] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794479, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.860032] env[61995]: DEBUG nova.compute.utils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.863574] env[61995]: INFO nova.compute.claims [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.868191] env[61995]: DEBUG nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 793.868571] env[61995]: DEBUG nova.network.neutron [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 793.924082] env[61995]: DEBUG nova.policy [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2388a3bd1587413285ff9f9ec07739b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9301d14211a1464eb740d9bb745b631a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 793.955408] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.006221] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794478, 'name': CreateVM_Task, 'duration_secs': 0.508101} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.006221] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 794.006221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.006221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.006221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 794.006221] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f88de0ac-76d0-458b-a4c3-9ebc775ec156 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.014362] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 794.014362] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f21fe9-9ea4-2f30-0e49-e8e3984a746a" [ 794.014362] env[61995]: _type = "Task" [ 794.014362] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.021089] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Releasing lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.021089] env[61995]: DEBUG nova.compute.manager [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received event network-vif-plugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.021089] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.021089] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.021089] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.021089] env[61995]: DEBUG nova.compute.manager [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] No waiting events found dispatching network-vif-plugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 794.021089] env[61995]: WARNING nova.compute.manager [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received unexpected event network-vif-plugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c for instance with vm_state building and task_state spawning. [ 794.021089] env[61995]: DEBUG nova.compute.manager [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received event network-changed-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.021674] env[61995]: DEBUG nova.compute.manager [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Refreshing instance network info cache due to event network-changed-6ee71b78-02c5-4f63-a764-d5f1bb61107c. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 794.022481] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Acquiring lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.022481] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Acquired lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.022481] env[61995]: DEBUG nova.network.neutron [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Refreshing network info cache for port 6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.024208] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528bb40b-d09b-d0f2-f9e5-d9a414e7563f, 'name': SearchDatastore_Task, 'duration_secs': 0.015141} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.030442] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.030605] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 0d34066f-5b8f-4bac-9b83-67d78987c4b3/0d34066f-5b8f-4bac-9b83-67d78987c4b3.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 794.031167] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-099b6e4e-0f81-422d-9a20-3fd77a9d22c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.039739] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f21fe9-9ea4-2f30-0e49-e8e3984a746a, 'name': SearchDatastore_Task, 'duration_secs': 0.010675} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.042175] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.042607] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.042909] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.043141] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.043376] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.043751] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 794.043751] env[61995]: value = "task-794480" [ 794.043751] env[61995]: _type = "Task" [ 794.043751] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.044274] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-239f0976-7d37-4ebd-bf3c-8b7e86066ce3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.055522] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.066661] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.066875] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 794.067708] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87c725dc-c784-49f9-8494-92224ea22c1f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.073570] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 794.073570] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525cbe38-b226-e021-ef9e-127a424e6c69" [ 794.073570] env[61995]: _type = "Task" [ 794.073570] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.082443] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525cbe38-b226-e021-ef9e-127a424e6c69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.178741] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794479, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.333829] env[61995]: DEBUG nova.network.neutron [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Successfully created port: 3393eaad-1226-4e95-866a-c9c4ad8858c3 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.370820] env[61995]: INFO nova.compute.resource_tracker [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating resource usage from migration e608fd92-962e-49ec-8fe1-0aeb1cf2b857 [ 794.376042] env[61995]: DEBUG nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.458261] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794475, 'name': ReconfigVM_Task, 'duration_secs': 0.634121} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.458754] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Reconfigured VM instance instance-00000030 to attach disk [datastore1] d056dadd-492d-4a4d-abc2-fc5f113f865b/d056dadd-492d-4a4d-abc2-fc5f113f865b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.459599] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21b60f53-6807-46c5-8d6b-c21c15d4ae06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.469537] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 794.469537] env[61995]: value = "task-794481" [ 794.469537] env[61995]: _type = "Task" [ 794.469537] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.480509] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794481, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.564465] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794480, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.587320] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525cbe38-b226-e021-ef9e-127a424e6c69, 'name': SearchDatastore_Task, 'duration_secs': 0.011847} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.591640] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f0d906-af0f-4164-a1ec-e685a9ed9007 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.598526] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 794.598526] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525cd367-2342-2ffc-b307-fb14f3fdf0a4" [ 794.598526] env[61995]: _type = "Task" [ 794.598526] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.611543] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525cd367-2342-2ffc-b307-fb14f3fdf0a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.678832] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794479, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.700714] env[61995]: DEBUG nova.compute.manager [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Received event network-changed-9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.700920] env[61995]: DEBUG nova.compute.manager [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Refreshing instance network info cache due to event network-changed-9116dcbf-d617-485e-afde-c82e908d5c15. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 794.701201] env[61995]: DEBUG oslo_concurrency.lockutils [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] Acquiring lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.701390] env[61995]: DEBUG oslo_concurrency.lockutils [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] Acquired lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.701504] env[61995]: DEBUG nova.network.neutron [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Refreshing network info cache for port 9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.800640] env[61995]: DEBUG nova.network.neutron [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updated VIF entry in instance network info cache for port 6ee71b78-02c5-4f63-a764-d5f1bb61107c. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 794.800954] env[61995]: DEBUG nova.network.neutron [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee71b78-02", "ovs_interfaceid": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.831765] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24d9558-aff6-47b5-b938-76c3d11dc73d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.840158] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e143a4-f22c-447c-9e84-83e214a2e7ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.872626] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa394ef4-e663-4bbb-97c0-833720a68eac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.883206] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b301a28e-496e-438f-a5bb-7d1e2c07f3f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.905916] env[61995]: DEBUG nova.compute.provider_tree [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.980056] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794481, 'name': Rename_Task, 'duration_secs': 0.443503} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.980426] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 794.980670] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0676d2f3-e7fe-40c1-8c66-e960c9a6c9b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.986905] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 794.986905] env[61995]: value = "task-794482" [ 794.986905] env[61995]: _type = "Task" [ 794.986905] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.995491] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.057700] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794480, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607839} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.058021] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 0d34066f-5b8f-4bac-9b83-67d78987c4b3/0d34066f-5b8f-4bac-9b83-67d78987c4b3.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 795.058533] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.058848] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a18fa64-3095-4a51-8eec-0963fac3b7bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.065654] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 795.065654] env[61995]: value = "task-794483" [ 795.065654] env[61995]: _type = "Task" [ 795.065654] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.073885] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.109025] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525cd367-2342-2ffc-b307-fb14f3fdf0a4, 'name': SearchDatastore_Task, 'duration_secs': 0.018821} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.109346] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.109395] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 795.109698] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5531218c-990c-4150-9cef-76ab4fddee0d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.116069] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 795.116069] env[61995]: value = "task-794484" [ 795.116069] env[61995]: _type = "Task" [ 795.116069] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.123565] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794484, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.181553] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794479, 'name': CreateSnapshot_Task, 'duration_secs': 1.250635} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.181895] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 795.182699] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c58aa8d-cfd8-4734-80bc-f6353d955a19 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.304042] env[61995]: DEBUG oslo_concurrency.lockutils [req-c397d3de-0e48-4de3-af85-aca4543f3340 req-f585cc2a-f85a-4217-800c-6e71937b9a6e service nova] Releasing lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.393350] env[61995]: DEBUG nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.410687] env[61995]: DEBUG nova.scheduler.client.report [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.416208] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.416462] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.416641] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.416839] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.416991] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.417157] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.417361] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.417550] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.417691] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.417856] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.418074] env[61995]: DEBUG nova.virt.hardware [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.418902] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf44e90-ed69-4a65-8050-63309f2cbec9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.427589] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f173297-f361-47b3-bfbe-2200f07ad578 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.485346] env[61995]: DEBUG nova.network.neutron [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Updated VIF entry in instance network info cache for port 9116dcbf-d617-485e-afde-c82e908d5c15. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 795.485765] env[61995]: DEBUG nova.network.neutron [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Updating instance_info_cache with network_info: [{"id": "9116dcbf-d617-485e-afde-c82e908d5c15", "address": "fa:16:3e:bc:1a:0f", "network": {"id": "bef08ad8-f37d-4b6a-a91c-e475beb866e7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1845121281-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "079a11ce311d40f9b8344fcce2e03683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9116dcbf-d6", "ovs_interfaceid": "9116dcbf-d617-485e-afde-c82e908d5c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.492334] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 795.492526] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 795.499087] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794482, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.315970] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 796.316681] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.966s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.316863] env[61995]: INFO nova.compute.manager [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Migrating [ 796.317101] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.317251] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.318534] env[61995]: DEBUG oslo_concurrency.lockutils [req-7bd847d0-c3c4-4217-be5c-002208e19f19 req-6116b4bc-f974-44d0-8ac4-ec8c194860ab service nova] Releasing lock "refresh_cache-eef1cf32-e71e-4696-8170-1023e3a5518d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.323178] env[61995]: DEBUG nova.compute.manager [req-306d32a2-f484-45d8-9d02-279fe0943a14 req-5b54154f-4cc6-4a97-bbd8-cb75b4f21f95 service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Received event network-vif-plugged-3393eaad-1226-4e95-866a-c9c4ad8858c3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 796.323468] env[61995]: DEBUG oslo_concurrency.lockutils [req-306d32a2-f484-45d8-9d02-279fe0943a14 req-5b54154f-4cc6-4a97-bbd8-cb75b4f21f95 service nova] Acquiring lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.323853] env[61995]: DEBUG oslo_concurrency.lockutils [req-306d32a2-f484-45d8-9d02-279fe0943a14 req-5b54154f-4cc6-4a97-bbd8-cb75b4f21f95 service nova] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.323853] env[61995]: DEBUG oslo_concurrency.lockutils [req-306d32a2-f484-45d8-9d02-279fe0943a14 req-5b54154f-4cc6-4a97-bbd8-cb75b4f21f95 service nova] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.324038] env[61995]: DEBUG nova.compute.manager [req-306d32a2-f484-45d8-9d02-279fe0943a14 req-5b54154f-4cc6-4a97-bbd8-cb75b4f21f95 service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] No waiting events found dispatching network-vif-plugged-3393eaad-1226-4e95-866a-c9c4ad8858c3 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.324216] env[61995]: WARNING nova.compute.manager [req-306d32a2-f484-45d8-9d02-279fe0943a14 req-5b54154f-4cc6-4a97-bbd8-cb75b4f21f95 service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Received unexpected event network-vif-plugged-3393eaad-1226-4e95-866a-c9c4ad8858c3 for instance with vm_state building and task_state spawning. [ 796.325139] env[61995]: DEBUG nova.network.neutron [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Successfully updated port: 3393eaad-1226-4e95-866a-c9c4ad8858c3 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.326574] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.326786] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.333334] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b9bd6dec-1dd3-4b27-b156-9b056946f52e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.337408] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.233s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.337408] env[61995]: DEBUG nova.objects.instance [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 796.343934] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.344052] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 796.344196] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 796.345727] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-823d2169-2bcb-4d5a-9677-cc4ac49a7e84" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.345875] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-823d2169-2bcb-4d5a-9677-cc4ac49a7e84" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.346661] env[61995]: DEBUG nova.network.neutron [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.359223] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794484, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.019593} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.359434] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065222} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.360909] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 796.363019] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.363019] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.363019] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 796.363019] env[61995]: value = "task-794485" [ 796.363019] env[61995]: _type = "Task" [ 796.363019] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.364809] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0546ebc9-d005-41b8-b49e-cbfc46a9d6f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.367474] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0bca74-bef9-4ecd-90d4-3d19b5e129de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.370929] env[61995]: DEBUG oslo_vmware.api [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794482, 'name': PowerOnVM_Task, 'duration_secs': 1.251848} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.373793] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 796.373990] env[61995]: INFO nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Took 10.86 seconds to spawn the instance on the hypervisor. [ 796.374186] env[61995]: DEBUG nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.375629] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929f3d01-88b1-4df6-ad30-e2724e875226 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.389988] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 796.389988] env[61995]: value = "task-794486" [ 796.389988] env[61995]: _type = "Task" [ 796.389988] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.398457] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 0d34066f-5b8f-4bac-9b83-67d78987c4b3/0d34066f-5b8f-4bac-9b83-67d78987c4b3.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.402656] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d9d7e6c-c4e7-49c5-9e6f-0d8e6ef46dd3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.416431] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794485, 'name': CloneVM_Task} progress is 12%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.430644] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794486, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.431382] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 796.431382] env[61995]: value = "task-794487" [ 796.431382] env[61995]: _type = "Task" [ 796.431382] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.439493] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794487, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.827212] env[61995]: INFO nova.compute.rpcapi [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 796.827789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.854170] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 796.855042] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 796.855042] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 796.855042] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 796.879359] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794485, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.887567] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.887567] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.887710] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 796.888100] env[61995]: DEBUG nova.objects.instance [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lazy-loading 'info_cache' on Instance uuid e53d10a3-1d16-498d-b4d8-abe7205410ec {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.889883] env[61995]: DEBUG nova.network.neutron [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.914094] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794486, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067184} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.914382] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.915324] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b138e2-88ad-4ea1-8180-2c9770a468ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.941171] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.945223] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c979991-2068-41c2-b98c-91dcb6e7d2eb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.959630] env[61995]: INFO nova.compute.manager [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Took 40.16 seconds to build instance. [ 796.969965] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794487, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.972347] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 796.972347] env[61995]: value = "task-794488" [ 796.972347] env[61995]: _type = "Task" [ 796.972347] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.981679] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794488, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.076043] env[61995]: DEBUG nova.network.neutron [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Updating instance_info_cache with network_info: [{"id": "3393eaad-1226-4e95-866a-c9c4ad8858c3", "address": "fa:16:3e:ad:3f:fa", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3393eaad-12", "ovs_interfaceid": "3393eaad-1226-4e95-866a-c9c4ad8858c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.349782] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.350085] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.350182] env[61995]: DEBUG nova.network.neutron [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 797.352624] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6425a057-7c59-4945-a2f0-2ac2c4ef40c0 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.353516] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.451s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.355859] env[61995]: INFO nova.compute.claims [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.382095] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794485, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.457187] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794487, 'name': ReconfigVM_Task, 'duration_secs': 0.822073} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.457488] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 0d34066f-5b8f-4bac-9b83-67d78987c4b3/0d34066f-5b8f-4bac-9b83-67d78987c4b3.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.458301] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db664e3d-6b82-4798-a825-dbc17cd161e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.465279] env[61995]: DEBUG oslo_concurrency.lockutils [None req-65e71bf9-9b89-49ac-9687-58bf65220ac7 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.764s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.465686] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 797.465686] env[61995]: value = "task-794489" [ 797.465686] env[61995]: _type = "Task" [ 797.465686] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.483555] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794488, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.486849] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794489, 'name': Rename_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.580126] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-823d2169-2bcb-4d5a-9677-cc4ac49a7e84" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.580126] env[61995]: DEBUG nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Instance network_info: |[{"id": "3393eaad-1226-4e95-866a-c9c4ad8858c3", "address": "fa:16:3e:ad:3f:fa", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3393eaad-12", "ovs_interfaceid": "3393eaad-1226-4e95-866a-c9c4ad8858c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 797.580484] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:3f:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3393eaad-1226-4e95-866a-c9c4ad8858c3', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.588158] env[61995]: DEBUG oslo.service.loopingcall [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.588399] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 797.588638] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-abc9bac3-9436-400f-9e0e-8ee128b1bbf3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.611593] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.611593] env[61995]: value = "task-794490" [ 797.611593] env[61995]: _type = "Task" [ 797.611593] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.620359] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794490, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.883019] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794485, 'name': CloneVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.974317] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.984057] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794489, 'name': Rename_Task, 'duration_secs': 0.188527} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.984573] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 797.985297] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-439d0d46-b56c-401d-b72c-52d61428c205 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.991106] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794488, 'name': ReconfigVM_Task, 'duration_secs': 0.851864} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.991843] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.992567] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce76f8ff-5fc8-495b-950b-4daf22ebedf8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.995915] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 797.995915] env[61995]: value = "task-794491" [ 797.995915] env[61995]: _type = "Task" [ 797.995915] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.999815] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "d056dadd-492d-4a4d-abc2-fc5f113f865b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.000074] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.000332] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "d056dadd-492d-4a4d-abc2-fc5f113f865b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.000638] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.000758] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.002931] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 798.002931] env[61995]: value = "task-794492" [ 798.002931] env[61995]: _type = "Task" [ 798.002931] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.003559] env[61995]: INFO nova.compute.manager [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Terminating instance [ 798.005538] env[61995]: DEBUG nova.compute.manager [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.005756] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 798.007062] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441498b9-ae62-430d-b6a8-1978ecf6dea1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.016073] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794491, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.022394] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 798.028069] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-177b36e3-3772-44e9-a7e5-d0f4f3e5c3bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.029755] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794492, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.035426] env[61995]: DEBUG oslo_vmware.api [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 798.035426] env[61995]: value = "task-794493" [ 798.035426] env[61995]: _type = "Task" [ 798.035426] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.044302] env[61995]: DEBUG oslo_vmware.api [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.121720] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794490, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.143632] env[61995]: DEBUG nova.network.neutron [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance_info_cache with network_info: [{"id": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "address": "fa:16:3e:3d:94:75", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap591f35b8-ba", "ovs_interfaceid": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.208632] env[61995]: DEBUG nova.compute.manager [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Received event network-changed-3393eaad-1226-4e95-866a-c9c4ad8858c3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 798.208780] env[61995]: DEBUG nova.compute.manager [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Refreshing instance network info cache due to event network-changed-3393eaad-1226-4e95-866a-c9c4ad8858c3. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 798.209016] env[61995]: DEBUG oslo_concurrency.lockutils [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] Acquiring lock "refresh_cache-823d2169-2bcb-4d5a-9677-cc4ac49a7e84" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.209250] env[61995]: DEBUG oslo_concurrency.lockutils [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] Acquired lock "refresh_cache-823d2169-2bcb-4d5a-9677-cc4ac49a7e84" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.209446] env[61995]: DEBUG nova.network.neutron [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Refreshing network info cache for port 3393eaad-1226-4e95-866a-c9c4ad8858c3 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.382478] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794485, 'name': CloneVM_Task, 'duration_secs': 1.574125} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.382816] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Created linked-clone VM from snapshot [ 798.383581] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea8d059-6c14-4dc9-b22c-6766c8c7e40c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.390932] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Uploading image 49cfe0c7-9c09-4a63-bd2f-322ba7cb789f {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 798.416534] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 798.416534] env[61995]: value = "vm-185293" [ 798.416534] env[61995]: _type = "VirtualMachine" [ 798.416534] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 798.416534] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f62d4cca-7a9e-4333-9301-39d4c9cf9e79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.427489] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lease: (returnval){ [ 798.427489] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5243c733-d6e8-8c78-45ee-9483dbb79a0e" [ 798.427489] env[61995]: _type = "HttpNfcLease" [ 798.427489] env[61995]: } obtained for exporting VM: (result){ [ 798.427489] env[61995]: value = "vm-185293" [ 798.427489] env[61995]: _type = "VirtualMachine" [ 798.427489] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 798.427489] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the lease: (returnval){ [ 798.427489] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5243c733-d6e8-8c78-45ee-9483dbb79a0e" [ 798.427489] env[61995]: _type = "HttpNfcLease" [ 798.427489] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 798.434913] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 798.434913] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5243c733-d6e8-8c78-45ee-9483dbb79a0e" [ 798.434913] env[61995]: _type = "HttpNfcLease" [ 798.434913] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 798.506198] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794491, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.509768] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.520507] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794492, 'name': Rename_Task, 'duration_secs': 0.184117} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.520507] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 798.523130] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01a2a423-2c52-4b1e-a44f-c0ba1996ffae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.529612] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 798.529612] env[61995]: value = "task-794495" [ 798.529612] env[61995]: _type = "Task" [ 798.529612] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.545253] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794495, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.550607] env[61995]: DEBUG oslo_vmware.api [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794493, 'name': PowerOffVM_Task, 'duration_secs': 0.212477} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.550899] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 798.551055] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 798.551308] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-825cf8db-7672-4af0-9bb7-be61dfe1185e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.624779] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794490, 'name': CreateVM_Task, 'duration_secs': 0.54359} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.627464] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 798.628476] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.628622] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.629268] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.629532] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2eb1a6b6-7748-4439-a401-a16de6add228 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.635063] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 798.635063] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52cc196a-f4a5-5d14-5e84-15a4d97877d3" [ 798.635063] env[61995]: _type = "Task" [ 798.635063] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.646097] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.653120] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52cc196a-f4a5-5d14-5e84-15a4d97877d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.705753] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.786200] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b111732-ca2c-4303-ab06-31cbfcb62e76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.794423] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb854fb-6775-4492-b0f1-ea43436e9e06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.827600] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b270004-3892-4358-acfb-782148965fbc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.835465] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f4fde3-ca4e-46cb-b20e-c11cf7a4af04 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.848826] env[61995]: DEBUG nova.compute.provider_tree [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.934378] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 798.934378] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5243c733-d6e8-8c78-45ee-9483dbb79a0e" [ 798.934378] env[61995]: _type = "HttpNfcLease" [ 798.934378] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 798.934690] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 798.934690] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5243c733-d6e8-8c78-45ee-9483dbb79a0e" [ 798.934690] env[61995]: _type = "HttpNfcLease" [ 798.934690] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 798.935400] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7cd768-7557-48cc-8c1f-6fc8c46efa30 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.943045] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b7b695-76db-a278-08a4-d6ed1cd0737c/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 798.943092] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b7b695-76db-a278-08a4-d6ed1cd0737c/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 799.014903] env[61995]: DEBUG oslo_vmware.api [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794491, 'name': PowerOnVM_Task, 'duration_secs': 0.525134} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.015305] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.015518] env[61995]: INFO nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Took 10.91 seconds to spawn the instance on the hypervisor. [ 799.016247] env[61995]: DEBUG nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.017032] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6f2ffb-3bdc-4c4c-bd18-7c78462bc655 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.038994] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794495, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.048396] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-54f100b0-dbd2-49f5-95b4-9bd3b028bede {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.071128] env[61995]: DEBUG nova.network.neutron [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Updated VIF entry in instance network info cache for port 3393eaad-1226-4e95-866a-c9c4ad8858c3. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.071400] env[61995]: DEBUG nova.network.neutron [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Updating instance_info_cache with network_info: [{"id": "3393eaad-1226-4e95-866a-c9c4ad8858c3", "address": "fa:16:3e:ad:3f:fa", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3393eaad-12", "ovs_interfaceid": "3393eaad-1226-4e95-866a-c9c4ad8858c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.145071] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52cc196a-f4a5-5d14-5e84-15a4d97877d3, 'name': SearchDatastore_Task, 'duration_secs': 0.012747} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.145286] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.145532] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.146198] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.146198] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.146198] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.146429] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f006240-5976-42f8-8ea6-9f652ef695de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.158039] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.158039] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 799.158693] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43a1ec9d-0341-4381-ac5d-bbe22560a053 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.165356] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 799.165356] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b3ee20-3a4e-8cbc-4b3b-f731d1c936a8" [ 799.165356] env[61995]: _type = "Task" [ 799.165356] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.175019] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b3ee20-3a4e-8cbc-4b3b-f731d1c936a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.208341] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.208341] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 799.208481] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.208672] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.208852] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.209084] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.209301] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.209488] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.210532] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 799.210532] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.353447] env[61995]: DEBUG nova.scheduler.client.report [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.534173] env[61995]: INFO nova.compute.manager [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Took 41.88 seconds to build instance. [ 799.545845] env[61995]: DEBUG oslo_vmware.api [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794495, 'name': PowerOnVM_Task, 'duration_secs': 0.854981} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.546247] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.546503] env[61995]: INFO nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Took 8.90 seconds to spawn the instance on the hypervisor. [ 799.546737] env[61995]: DEBUG nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.547649] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05205c8-666c-4bd9-8a62-93aaab9edabc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.574655] env[61995]: DEBUG oslo_concurrency.lockutils [req-9c9ac616-f7e7-4053-903a-1353bc27aef8 req-e17e6c92-5611-4f0e-9b27-c3835c4bf72f service nova] Releasing lock "refresh_cache-823d2169-2bcb-4d5a-9677-cc4ac49a7e84" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.678805] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b3ee20-3a4e-8cbc-4b3b-f731d1c936a8, 'name': SearchDatastore_Task, 'duration_secs': 0.012146} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.680073] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fd2e096-3962-4a04-a1b7-fb9511e04e81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.687266] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 799.687266] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52fc72da-c2ad-75d6-bdf1-d8ee1ccda6f9" [ 799.687266] env[61995]: _type = "Task" [ 799.687266] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.697795] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fc72da-c2ad-75d6-bdf1-d8ee1ccda6f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.713021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.860826] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.861484] env[61995]: DEBUG nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.865378] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.825s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.865614] env[61995]: DEBUG nova.objects.instance [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lazy-loading 'resources' on Instance uuid ecb72668-caa4-4117-9f8d-e0c6e4831f3d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.037811] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c3b2f0a8-461b-465c-8c50-a664e9d596f7 tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.222s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.066379] env[61995]: INFO nova.compute.manager [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Took 37.30 seconds to build instance. [ 800.170486] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155fc2f6-6f63-4605-b5f1-f0004fc850e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.202292] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance '4edbf65d-eaca-47e1-b6bf-03ccb908b52e' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 800.215801] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 800.216294] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 800.216884] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Deleting the datastore file [datastore1] d056dadd-492d-4a4d-abc2-fc5f113f865b {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 800.217748] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de878c46-8398-4357-82b6-90a7348a015a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.231866] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fc72da-c2ad-75d6-bdf1-d8ee1ccda6f9, 'name': SearchDatastore_Task, 'duration_secs': 0.012097} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.234662] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.235134] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 800.235767] env[61995]: DEBUG oslo_vmware.api [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for the task: (returnval){ [ 800.235767] env[61995]: value = "task-794497" [ 800.235767] env[61995]: _type = "Task" [ 800.235767] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.236387] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2be11aec-c639-498a-81f0-bafe450b84c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.254040] env[61995]: DEBUG oslo_vmware.api [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.258055] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 800.258055] env[61995]: value = "task-794498" [ 800.258055] env[61995]: _type = "Task" [ 800.258055] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.273097] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.367017] env[61995]: DEBUG nova.compute.utils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.368829] env[61995]: DEBUG nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.369614] env[61995]: DEBUG nova.network.neutron [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 800.457752] env[61995]: DEBUG nova.policy [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f27aa509c254e2fa6a5d833b53f50e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4042c97b25a24bf0a1ebe23e958f95ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 800.541993] env[61995]: DEBUG nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.570354] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a0bffc71-3b7c-44d8-bf2a-11be6fdf1546 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.925s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.720649] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 800.721202] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ecc7372-7e8a-4542-a5fb-22c273392f3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.733830] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 800.733830] env[61995]: value = "task-794499" [ 800.733830] env[61995]: _type = "Task" [ 800.733830] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.751580] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794499, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.764312] env[61995]: DEBUG oslo_vmware.api [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Task: {'id': task-794497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.292846} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.769834] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 800.770298] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 800.770688] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 800.771430] env[61995]: INFO nova.compute.manager [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Took 2.77 seconds to destroy the instance on the hypervisor. [ 800.771582] env[61995]: DEBUG oslo.service.loopingcall [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.776115] env[61995]: DEBUG nova.compute.manager [-] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.776297] env[61995]: DEBUG nova.network.neutron [-] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 800.788623] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794498, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.877242] env[61995]: DEBUG nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.935784] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645fe90d-dc07-4d31-b21f-d79fbf918895 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.945992] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327ae818-dbb9-407b-a29a-8b81911a7b9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.984301] env[61995]: DEBUG nova.network.neutron [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Successfully created port: b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.987317] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21183ee5-fe21-42c7-875c-6977cd1d7d77 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.996760] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b574b6f4-3f70-4e0f-9528-eb1421668854 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.016328] env[61995]: DEBUG nova.compute.provider_tree [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.076217] env[61995]: DEBUG nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.077805] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.246137] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794499, 'name': PowerOffVM_Task, 'duration_secs': 0.43227} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.247186] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 801.247243] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance '4edbf65d-eaca-47e1-b6bf-03ccb908b52e' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 801.278995] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672758} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.279372] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 801.279625] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.279942] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81a49d23-4f01-4129-868f-f056a6f5a4d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.289433] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 801.289433] env[61995]: value = "task-794500" [ 801.289433] env[61995]: _type = "Task" [ 801.289433] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.299612] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794500, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.523029] env[61995]: DEBUG nova.scheduler.client.report [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.594488] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.702835] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.703606] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.749582] env[61995]: DEBUG nova.compute.manager [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 801.749763] env[61995]: DEBUG nova.compute.manager [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing instance network info cache due to event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 801.749990] env[61995]: DEBUG oslo_concurrency.lockutils [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] Acquiring lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.750154] env[61995]: DEBUG oslo_concurrency.lockutils [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] Acquired lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.750341] env[61995]: DEBUG nova.network.neutron [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 801.753417] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.753640] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.753802] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.753989] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.754169] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.754346] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.754557] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.754739] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.754932] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.755166] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.755321] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.763572] env[61995]: DEBUG nova.network.neutron [-] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.765669] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39575089-bed2-4b90-80e8-a50b30da9a4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.779836] env[61995]: INFO nova.compute.manager [-] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Took 1.00 seconds to deallocate network for instance. [ 801.788426] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 801.788426] env[61995]: value = "task-794501" [ 801.788426] env[61995]: _type = "Task" [ 801.788426] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.803401] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794501, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.806831] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794500, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114701} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.807158] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.807967] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a71698f-f9d1-434b-aaca-0317509cbef0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.835258] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.838833] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07e3d5d1-a43c-4bf2-9dd4-e03b62f99bd3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.860566] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 801.860566] env[61995]: value = "task-794502" [ 801.860566] env[61995]: _type = "Task" [ 801.860566] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.869983] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794502, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.891430] env[61995]: DEBUG nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.919823] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.920113] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.920300] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.920700] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.920700] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.920787] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.921010] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.921260] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.921444] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.921624] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.921825] env[61995]: DEBUG nova.virt.hardware [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.923480] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a289836b-4d43-4a43-a111-4e188bfca700 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.938177] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345079db-c0c4-4f06-88c2-db70051ec8b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.028187] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.163s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.032039] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.928s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.036144] env[61995]: INFO nova.compute.claims [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.080822] env[61995]: INFO nova.scheduler.client.report [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleted allocations for instance ecb72668-caa4-4117-9f8d-e0c6e4831f3d [ 802.147828] env[61995]: DEBUG nova.network.neutron [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updated VIF entry in instance network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 802.148328] env[61995]: DEBUG nova.network.neutron [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.287765] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.298661] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794501, 'name': ReconfigVM_Task, 'duration_secs': 0.292634} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.299016] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance '4edbf65d-eaca-47e1-b6bf-03ccb908b52e' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 802.370317] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794502, 'name': ReconfigVM_Task, 'duration_secs': 0.43423} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.370630] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.371266] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c07d04b7-689b-40d8-ab9b-c86ba13fed3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.377505] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 802.377505] env[61995]: value = "task-794503" [ 802.377505] env[61995]: _type = "Task" [ 802.377505] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.389393] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794503, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.591589] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b9952cc-8f48-4cb5-a0ca-c711c369a971 tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "ecb72668-caa4-4117-9f8d-e0c6e4831f3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.316s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.651886] env[61995]: DEBUG oslo_concurrency.lockutils [req-b9633b2a-bc42-441e-92c4-7b5a9e2ae876 req-3bab36b8-8024-4e40-bb41-9d0989dea64e service nova] Releasing lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.806613] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.807059] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.807346] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.807671] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.807948] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.808979] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.808979] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.808979] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.809346] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.809550] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.810262] env[61995]: DEBUG nova.virt.hardware [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.824234] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Reconfiguring VM instance instance-00000029 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 802.824500] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29945d2c-96a2-4a12-8afd-8a93e8b7b9ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.852409] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 802.852409] env[61995]: value = "task-794504" [ 802.852409] env[61995]: _type = "Task" [ 802.852409] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.861199] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794504, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.888426] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794503, 'name': Rename_Task, 'duration_secs': 0.297621} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.888810] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 802.889116] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b47bd177-4b6c-42aa-994f-bd279268a3de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.895558] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 802.895558] env[61995]: value = "task-794505" [ 802.895558] env[61995]: _type = "Task" [ 802.895558] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.904419] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.033165] env[61995]: DEBUG nova.compute.manager [req-90605f38-756c-4292-bcc7-158c5aea46af req-70cf502a-4a20-46e1-9a9d-6550cb823246 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Received event network-vif-plugged-b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.033611] env[61995]: DEBUG oslo_concurrency.lockutils [req-90605f38-756c-4292-bcc7-158c5aea46af req-70cf502a-4a20-46e1-9a9d-6550cb823246 service nova] Acquiring lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.033660] env[61995]: DEBUG oslo_concurrency.lockutils [req-90605f38-756c-4292-bcc7-158c5aea46af req-70cf502a-4a20-46e1-9a9d-6550cb823246 service nova] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.033827] env[61995]: DEBUG oslo_concurrency.lockutils [req-90605f38-756c-4292-bcc7-158c5aea46af req-70cf502a-4a20-46e1-9a9d-6550cb823246 service nova] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.034086] env[61995]: DEBUG nova.compute.manager [req-90605f38-756c-4292-bcc7-158c5aea46af req-70cf502a-4a20-46e1-9a9d-6550cb823246 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] No waiting events found dispatching network-vif-plugged-b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.034305] env[61995]: WARNING nova.compute.manager [req-90605f38-756c-4292-bcc7-158c5aea46af req-70cf502a-4a20-46e1-9a9d-6550cb823246 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Received unexpected event network-vif-plugged-b970a5e5-c432-4698-bc92-c64f8b7e3091 for instance with vm_state building and task_state spawning. [ 803.155802] env[61995]: DEBUG nova.network.neutron [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Successfully updated port: b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.363760] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794504, 'name': ReconfigVM_Task, 'duration_secs': 0.227509} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.366618] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Reconfigured VM instance instance-00000029 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 803.368591] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600575d4-169c-4ca7-abfc-b05e0937c5cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.391887] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 4edbf65d-eaca-47e1-b6bf-03ccb908b52e/4edbf65d-eaca-47e1-b6bf-03ccb908b52e.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.395065] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4a4294a-5025-4f83-8686-9fa9ae0d89ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.421723] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.423375] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 803.423375] env[61995]: value = "task-794506" [ 803.423375] env[61995]: _type = "Task" [ 803.423375] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.434096] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794506, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.532497] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96881ae2-4df2-4063-a8bc-26daa14c2123 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.540937] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7056af-4d93-4686-8b2c-b7c9113273d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.575072] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76009461-2691-4165-861e-48c52f7c32db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.584318] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d8017c-2086-470b-a8f3-01022e730ed1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.600989] env[61995]: DEBUG nova.compute.provider_tree [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.658901] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.659193] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.659265] env[61995]: DEBUG nova.network.neutron [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.892310] env[61995]: DEBUG nova.compute.manager [req-62831e7c-188d-482f-99b4-8cbe018140af req-44040cad-2273-42e7-a435-d7d455a9e6e4 service nova] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Received event network-vif-deleted-647ca67a-7528-4c2a-8500-6dac05d4ed89 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.924052] env[61995]: DEBUG oslo_vmware.api [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794505, 'name': PowerOnVM_Task, 'duration_secs': 0.716414} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.928767] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 803.929015] env[61995]: INFO nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Took 8.54 seconds to spawn the instance on the hypervisor. [ 803.929310] env[61995]: DEBUG nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.930697] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c35a59c-e80b-4be4-a715-7bd916c9d2fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.939931] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794506, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.977707] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "55555deb-9883-483b-a901-42b1a66c60d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.978074] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "55555deb-9883-483b-a901-42b1a66c60d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.978346] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "55555deb-9883-483b-a901-42b1a66c60d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.978644] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "55555deb-9883-483b-a901-42b1a66c60d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.979280] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "55555deb-9883-483b-a901-42b1a66c60d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.981870] env[61995]: INFO nova.compute.manager [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Terminating instance [ 803.983572] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "refresh_cache-55555deb-9883-483b-a901-42b1a66c60d6" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.983755] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquired lock "refresh_cache-55555deb-9883-483b-a901-42b1a66c60d6" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.984017] env[61995]: DEBUG nova.network.neutron [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.104151] env[61995]: DEBUG nova.scheduler.client.report [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.199661] env[61995]: DEBUG nova.network.neutron [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.366979] env[61995]: DEBUG nova.network.neutron [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Updating instance_info_cache with network_info: [{"id": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "address": "fa:16:3e:1e:e0:ed", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb970a5e5-c4", "ovs_interfaceid": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.432602] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794506, 'name': ReconfigVM_Task, 'duration_secs': 0.559097} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.432897] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 4edbf65d-eaca-47e1-b6bf-03ccb908b52e/4edbf65d-eaca-47e1-b6bf-03ccb908b52e.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.433236] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance '4edbf65d-eaca-47e1-b6bf-03ccb908b52e' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 804.455957] env[61995]: INFO nova.compute.manager [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Took 38.39 seconds to build instance. [ 804.508227] env[61995]: DEBUG nova.network.neutron [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.589222] env[61995]: DEBUG nova.network.neutron [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.608839] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.609549] env[61995]: DEBUG nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.612942] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.468s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.613215] env[61995]: DEBUG nova.objects.instance [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lazy-loading 'resources' on Instance uuid 2da17a43-2260-4a82-9ed3-eafd25bb9a6a {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.869393] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.869786] env[61995]: DEBUG nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Instance network_info: |[{"id": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "address": "fa:16:3e:1e:e0:ed", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb970a5e5-c4", "ovs_interfaceid": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.870211] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:e0:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3291573-fad8-48cc-a965-c3554e7cee4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b970a5e5-c432-4698-bc92-c64f8b7e3091', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.878016] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Creating folder: Project (4042c97b25a24bf0a1ebe23e958f95ac). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.879247] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b01f2694-abc6-47ae-8a12-7e9b3c6b7577 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.890437] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Created folder: Project (4042c97b25a24bf0a1ebe23e958f95ac) in parent group-v185203. [ 804.890679] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Creating folder: Instances. Parent ref: group-v185295. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.890930] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d65a4220-1b70-4892-b68c-da9a5c47c006 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.900279] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Created folder: Instances in parent group-v185295. [ 804.900557] env[61995]: DEBUG oslo.service.loopingcall [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.900754] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 804.900971] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-915926e9-5b23-4be1-bde4-b1eff1f9496c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.920971] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.920971] env[61995]: value = "task-794509" [ 804.920971] env[61995]: _type = "Task" [ 804.920971] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.929517] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794509, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.939922] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390f54d5-ce50-4eab-98ac-f6662c4d6427 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.961288] env[61995]: DEBUG oslo_concurrency.lockutils [None req-626b4e1e-1813-469d-aa3a-e80957a28f71 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.506s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.963215] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a352ad-1068-4e00-bff1-cd40960f9f10 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.983259] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance '4edbf65d-eaca-47e1-b6bf-03ccb908b52e' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 805.095152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Releasing lock "refresh_cache-55555deb-9883-483b-a901-42b1a66c60d6" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.095152] env[61995]: DEBUG nova.compute.manager [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.095152] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 805.095152] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ca2418-e2c9-472a-86ad-7ad2d56eeaae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.102778] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 805.103120] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4e1dd78-16d3-442f-9e9b-ceb280b3343a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.109930] env[61995]: DEBUG oslo_vmware.api [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 805.109930] env[61995]: value = "task-794510" [ 805.109930] env[61995]: _type = "Task" [ 805.109930] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.116959] env[61995]: DEBUG nova.compute.utils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.124876] env[61995]: DEBUG nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.124990] env[61995]: DEBUG nova.network.neutron [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 805.133180] env[61995]: DEBUG oslo_vmware.api [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.135097] env[61995]: DEBUG nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.248497] env[61995]: DEBUG nova.compute.manager [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 805.248954] env[61995]: DEBUG nova.compute.manager [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing instance network info cache due to event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 805.249387] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Acquiring lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.249823] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Acquired lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.250183] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.255195] env[61995]: DEBUG nova.policy [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3efdbd0f8e0340dc8e4bfbf672d62f0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83064cbae17c429d8d084837635486da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 805.357304] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.357576] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.357793] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.358044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.358284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.363825] env[61995]: INFO nova.compute.manager [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Terminating instance [ 805.367723] env[61995]: DEBUG nova.compute.manager [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.368052] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 805.369231] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1e7f5c-008b-4722-b0f5-c3db0e64202d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.378273] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 805.378541] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9be208fb-5126-445d-b14e-23e0edd91df3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.384893] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 805.384893] env[61995]: value = "task-794511" [ 805.384893] env[61995]: _type = "Task" [ 805.384893] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.394836] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.431545] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794509, 'name': CreateVM_Task, 'duration_secs': 0.457095} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.431717] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.432459] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.432633] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.432965] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.433566] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4bea0e8-230c-4a93-bbf1-944e9b980f5a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.439531] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 805.439531] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5213036f-5185-9b9c-1188-80626211ffbe" [ 805.439531] env[61995]: _type = "Task" [ 805.439531] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.447880] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5213036f-5185-9b9c-1188-80626211ffbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.467824] env[61995]: DEBUG nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.567179] env[61995]: DEBUG nova.network.neutron [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Port 591f35b8-baf5-4da1-b479-dc99ae2f7b75 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 805.616959] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b107b9a-ef0d-4e3a-a31d-5eea2bbcb9a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.627854] env[61995]: DEBUG oslo_vmware.api [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794510, 'name': PowerOffVM_Task, 'duration_secs': 0.136621} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.630267] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 805.630494] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 805.630796] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b66ed4e-0071-413f-9de5-bfaa6f1e7de7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.633344] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2296a56-2150-4ba7-a82e-c4a7d59c9948 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.672021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d597c2b-4b83-4c2c-a2d3-ab87706f3b15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.673613] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 805.673822] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 805.674015] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleting the datastore file [datastore1] 55555deb-9883-483b-a901-42b1a66c60d6 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.674482] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df4980e9-eb68-4a3e-a458-466753d2cc18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.685923] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec326e56-ee08-4149-aad1-ced59ed4e754 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.690214] env[61995]: DEBUG oslo_vmware.api [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for the task: (returnval){ [ 805.690214] env[61995]: value = "task-794513" [ 805.690214] env[61995]: _type = "Task" [ 805.690214] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.701297] env[61995]: DEBUG nova.compute.provider_tree [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.708853] env[61995]: DEBUG oslo_vmware.api [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.839509] env[61995]: DEBUG nova.network.neutron [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Successfully created port: ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.894357] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.950787] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5213036f-5185-9b9c-1188-80626211ffbe, 'name': SearchDatastore_Task, 'duration_secs': 0.016495} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.950973] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.951247] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.951494] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.951694] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.951841] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.952131] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f938bcdd-5d92-4ab1-99fd-a26e7a214516 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.961129] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.961330] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 805.962083] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5957c1ba-e773-4689-9913-4d44a410883a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.968749] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 805.968749] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f40c4a-7162-13d0-796c-91dc61df9875" [ 805.968749] env[61995]: _type = "Task" [ 805.968749] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.986103] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f40c4a-7162-13d0-796c-91dc61df9875, 'name': SearchDatastore_Task} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.986907] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2f02fbc-1f44-4539-a7ab-a8929c144ce1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.992963] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 805.992963] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523fb53a-a302-07df-5ec2-764d79ba79e4" [ 805.992963] env[61995]: _type = "Task" [ 805.992963] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.997357] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.003633] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523fb53a-a302-07df-5ec2-764d79ba79e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.104342] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updated VIF entry in instance network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.104853] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.171627] env[61995]: DEBUG nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.203149] env[61995]: DEBUG oslo_vmware.api [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Task: {'id': task-794513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188256} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.203418] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.203602] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 806.204197] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 806.204197] env[61995]: INFO nova.compute.manager [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 806.204282] env[61995]: DEBUG oslo.service.loopingcall [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.204705] env[61995]: DEBUG nova.compute.manager [-] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.204705] env[61995]: DEBUG nova.network.neutron [-] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.207828] env[61995]: DEBUG nova.scheduler.client.report [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.212628] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.213377] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.213377] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.213377] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.213548] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.213548] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.213921] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.213921] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.214069] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.218083] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.218083] env[61995]: DEBUG nova.virt.hardware [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.218083] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd3fede-76c5-470e-ac15-a372e4cdc2c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.226657] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899d484d-8e31-4312-94ea-98754a5e7c79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.245973] env[61995]: DEBUG nova.compute.manager [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Received event network-changed-2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.247035] env[61995]: DEBUG nova.compute.manager [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Refreshing instance network info cache due to event network-changed-2c08d25b-1e7a-4f2d-8715-0e64edc11b01. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 806.247035] env[61995]: DEBUG oslo_concurrency.lockutils [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] Acquiring lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.247035] env[61995]: DEBUG oslo_concurrency.lockutils [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] Acquired lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.247035] env[61995]: DEBUG nova.network.neutron [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Refreshing network info cache for port 2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.250874] env[61995]: DEBUG nova.network.neutron [-] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.395756] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.506562] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523fb53a-a302-07df-5ec2-764d79ba79e4, 'name': SearchDatastore_Task, 'duration_secs': 0.013168} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.506995] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.507532] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] dc7077ac-d3fd-4e84-867a-d86328b32f5d/dc7077ac-d3fd-4e84-867a-d86328b32f5d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 806.507757] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6afd3669-f207-4450-9650-320109471d5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.515473] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 806.515473] env[61995]: value = "task-794514" [ 806.515473] env[61995]: _type = "Task" [ 806.515473] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.524770] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.598356] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.602025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.602025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.607714] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Releasing lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.607996] env[61995]: DEBUG nova.compute.manager [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Received event network-changed-2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.608204] env[61995]: DEBUG nova.compute.manager [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Refreshing instance network info cache due to event network-changed-2c08d25b-1e7a-4f2d-8715-0e64edc11b01. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 806.608402] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Acquiring lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.721324] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.108s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.723770] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.649s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.725274] env[61995]: INFO nova.compute.claims [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.747764] env[61995]: INFO nova.scheduler.client.report [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Deleted allocations for instance 2da17a43-2260-4a82-9ed3-eafd25bb9a6a [ 806.755707] env[61995]: DEBUG nova.network.neutron [-] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.903716] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794511, 'name': PowerOffVM_Task, 'duration_secs': 1.321065} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.906430] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 806.906626] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 806.906931] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78f44d22-97dd-465a-91a4-7f85434a6471 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.012567] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 807.012819] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 807.013050] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Deleting the datastore file [datastore1] 0d34066f-5b8f-4bac-9b83-67d78987c4b3 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.014290] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13d82fe8-cd95-4d29-96a4-403453c6295b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.021410] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 807.021410] env[61995]: value = "task-794516" [ 807.021410] env[61995]: _type = "Task" [ 807.021410] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.028599] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794514, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.034562] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794516, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.109989] env[61995]: INFO nova.compute.manager [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Rebuilding instance [ 807.128839] env[61995]: DEBUG nova.network.neutron [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Updated VIF entry in instance network info cache for port 2c08d25b-1e7a-4f2d-8715-0e64edc11b01. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 807.129351] env[61995]: DEBUG nova.network.neutron [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Updating instance_info_cache with network_info: [{"id": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "address": "fa:16:3e:c1:25:58", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c08d25b-1e", "ovs_interfaceid": "2c08d25b-1e7a-4f2d-8715-0e64edc11b01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.154660] env[61995]: DEBUG nova.compute.manager [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.155538] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7223bc-0fda-4751-9021-3e1810686fb8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.258070] env[61995]: INFO nova.compute.manager [-] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Took 1.05 seconds to deallocate network for instance. [ 807.260310] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dccf766a-f95c-429f-a49c-c1c5be61cd2f tempest-ServerPasswordTestJSON-468588164 tempest-ServerPasswordTestJSON-468588164-project-member] Lock "2da17a43-2260-4a82-9ed3-eafd25bb9a6a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.537s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.459372] env[61995]: DEBUG nova.compute.manager [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 807.459580] env[61995]: DEBUG nova.compute.manager [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing instance network info cache due to event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 807.459807] env[61995]: DEBUG oslo_concurrency.lockutils [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] Acquiring lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.459954] env[61995]: DEBUG oslo_concurrency.lockutils [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] Acquired lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.461417] env[61995]: DEBUG nova.network.neutron [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 807.525957] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592469} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.529225] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] dc7077ac-d3fd-4e84-867a-d86328b32f5d/dc7077ac-d3fd-4e84-867a-d86328b32f5d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.529510] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.530127] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ecbdc106-a0e6-413c-a8c3-a3afd94f0071 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.539675] env[61995]: DEBUG oslo_vmware.api [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794516, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.358336} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.540779] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 807.541009] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 807.541209] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 807.541391] env[61995]: INFO nova.compute.manager [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Took 2.17 seconds to destroy the instance on the hypervisor. [ 807.541642] env[61995]: DEBUG oslo.service.loopingcall [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.541913] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 807.541913] env[61995]: value = "task-794517" [ 807.541913] env[61995]: _type = "Task" [ 807.541913] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.542125] env[61995]: DEBUG nova.compute.manager [-] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.542225] env[61995]: DEBUG nova.network.neutron [-] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 807.554188] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794517, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.607883] env[61995]: DEBUG nova.network.neutron [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Successfully updated port: ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.632092] env[61995]: DEBUG oslo_concurrency.lockutils [req-c9a634c7-959a-441a-8193-da82fa4fc4d5 req-eded74f4-b1ed-4ec5-9223-fdee1aae84fe service nova] Releasing lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.632856] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Acquired lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.633072] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Refreshing network info cache for port 2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 807.667549] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.667691] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-987c0c52-3309-4412-88b0-cf20d1af8eef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.679026] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 807.679026] env[61995]: value = "task-794518" [ 807.679026] env[61995]: _type = "Task" [ 807.679026] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.686249] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794518, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.696019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.696256] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.696460] env[61995]: DEBUG nova.network.neutron [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.765328] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.056854] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794517, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156478} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.057251] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.057934] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd37eacf-8267-4c9b-952c-e07fad176fd1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.083837] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] dc7077ac-d3fd-4e84-867a-d86328b32f5d/dc7077ac-d3fd-4e84-867a-d86328b32f5d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.086801] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea0f12db-9df0-4368-bb43-d14822092a12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.110986] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.111316] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.111446] env[61995]: DEBUG nova.network.neutron [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 808.118739] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 808.118739] env[61995]: value = "task-794519" [ 808.118739] env[61995]: _type = "Task" [ 808.118739] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.128766] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.167064] env[61995]: INFO nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Port 2c08d25b-1e7a-4f2d-8715-0e64edc11b01 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 808.167296] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.188440] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794518, 'name': PowerOffVM_Task, 'duration_secs': 0.245831} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.191154] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 808.191392] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 808.192713] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc500365-f0c3-4637-9cec-101a27dd4b7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.200066] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 808.200066] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b81bf7e-8b97-4303-b588-d6b8a1a40ab8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.204494] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029864ce-4209-4d99-b072-6327cb69b446 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.211831] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc85d465-fc4d-4d96-b08c-f8bdcea87c2c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.247067] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2488c69a-2272-4621-ba1a-0c2de39a8889 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.255163] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a41ce1-8a4e-4be2-bd0f-0e5302568939 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.273493] env[61995]: DEBUG nova.compute.provider_tree [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.487045] env[61995]: DEBUG nova.network.neutron [-] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.630975] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.646902] env[61995]: DEBUG nova.network.neutron [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updated VIF entry in instance network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 808.647293] env[61995]: DEBUG nova.network.neutron [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.670319] env[61995]: DEBUG nova.network.neutron [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.672592] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Releasing lock "refresh_cache-0d34066f-5b8f-4bac-9b83-67d78987c4b3" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.672884] env[61995]: DEBUG nova.compute.manager [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Received event network-changed-b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.673110] env[61995]: DEBUG nova.compute.manager [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Refreshing instance network info cache due to event network-changed-b970a5e5-c432-4698-bc92-c64f8b7e3091. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 808.673356] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Acquiring lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.673507] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Acquired lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.673674] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Refreshing network info cache for port b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.698745] env[61995]: DEBUG nova.compute.manager [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Received event network-vif-deleted-2c08d25b-1e7a-4f2d-8715-0e64edc11b01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.698894] env[61995]: DEBUG nova.compute.manager [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.699474] env[61995]: DEBUG nova.compute.manager [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing instance network info cache due to event network-changed-52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 808.699722] env[61995]: DEBUG oslo_concurrency.lockutils [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] Acquiring lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.770402] env[61995]: DEBUG nova.network.neutron [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance_info_cache with network_info: [{"id": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "address": "fa:16:3e:3d:94:75", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap591f35b8-ba", "ovs_interfaceid": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.777239] env[61995]: DEBUG nova.scheduler.client.report [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.988988] env[61995]: INFO nova.compute.manager [-] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Took 1.45 seconds to deallocate network for instance. [ 809.134281] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794519, 'name': ReconfigVM_Task, 'duration_secs': 0.990559} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.134894] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Reconfigured VM instance instance-00000034 to attach disk [datastore2] dc7077ac-d3fd-4e84-867a-d86328b32f5d/dc7077ac-d3fd-4e84-867a-d86328b32f5d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.135824] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c02c7a7a-9226-4a25-a613-c54cbe6ccb07 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.143543] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 809.143543] env[61995]: value = "task-794521" [ 809.143543] env[61995]: _type = "Task" [ 809.143543] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.159017] env[61995]: DEBUG oslo_concurrency.lockutils [req-f2982d2b-e380-4246-bc44-09418fdf42f5 req-b0f8a35d-1c3a-41b6-b742-99f870e3806a service nova] Releasing lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.160369] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794521, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.160654] env[61995]: DEBUG oslo_concurrency.lockutils [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] Acquired lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.160850] env[61995]: DEBUG nova.network.neutron [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Refreshing network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.167161] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 809.167374] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 809.167713] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleting the datastore file [datastore2] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.167820] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89fe6f08-2647-47f7-a233-91be2239cf9d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.174364] env[61995]: DEBUG nova.network.neutron [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updating instance_info_cache with network_info: [{"id": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "address": "fa:16:3e:d6:67:26", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcee8fc-e0", "ovs_interfaceid": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.182841] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 809.182841] env[61995]: value = "task-794522" [ 809.182841] env[61995]: _type = "Task" [ 809.182841] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.192572] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.273449] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.288959] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.288959] env[61995]: DEBUG nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.296698] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.179s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.298372] env[61995]: INFO nova.compute.claims [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.459529] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Updated VIF entry in instance network info cache for port b970a5e5-c432-4698-bc92-c64f8b7e3091. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.459920] env[61995]: DEBUG nova.network.neutron [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Updating instance_info_cache with network_info: [{"id": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "address": "fa:16:3e:1e:e0:ed", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb970a5e5-c4", "ovs_interfaceid": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.492194] env[61995]: DEBUG nova.compute.manager [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Received event network-vif-plugged-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 809.492581] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] Acquiring lock "b40b1866-ced3-40ef-9ab7-30d58b75f288-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.492802] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.493053] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.493289] env[61995]: DEBUG nova.compute.manager [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] No waiting events found dispatching network-vif-plugged-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 809.493505] env[61995]: WARNING nova.compute.manager [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Received unexpected event network-vif-plugged-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f for instance with vm_state building and task_state spawning. [ 809.493685] env[61995]: DEBUG nova.compute.manager [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Received event network-changed-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 809.493876] env[61995]: DEBUG nova.compute.manager [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Refreshing instance network info cache due to event network-changed-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 809.494076] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] Acquiring lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.496875] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.657051] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794521, 'name': Rename_Task, 'duration_secs': 0.194624} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.657051] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 809.657051] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c53db963-086e-4cfe-9231-bb0d86eb84cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.664717] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 809.664717] env[61995]: value = "task-794523" [ 809.664717] env[61995]: _type = "Task" [ 809.664717] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.680142] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.680633] env[61995]: DEBUG nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Instance network_info: |[{"id": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "address": "fa:16:3e:d6:67:26", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcee8fc-e0", "ovs_interfaceid": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.681672] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794523, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.682076] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] Acquired lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.682938] env[61995]: DEBUG nova.network.neutron [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Refreshing network info cache for port ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.685586] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:67:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.693976] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating folder: Project (83064cbae17c429d8d084837635486da). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.696247] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c1340ee6-3146-4550-9af5-19f11ac89ace {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.709974] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237663} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.711375] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.711579] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 809.711828] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 809.715849] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created folder: Project (83064cbae17c429d8d084837635486da) in parent group-v185203. [ 809.716130] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating folder: Instances. Parent ref: group-v185298. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.716453] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f57c0967-c737-4c35-b944-93815aeca2c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.727889] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created folder: Instances in parent group-v185298. [ 809.728266] env[61995]: DEBUG oslo.service.loopingcall [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.728513] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 809.728801] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee8e2625-71e1-4e78-8b6e-dceb715483f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.750381] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.750381] env[61995]: value = "task-794526" [ 809.750381] env[61995]: _type = "Task" [ 809.750381] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.759096] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794526, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.800018] env[61995]: DEBUG nova.compute.utils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.800295] env[61995]: DEBUG nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.800295] env[61995]: DEBUG nova.network.neutron [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.817866] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a4517f-e4bd-44f9-ac6e-17fe72d3955b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.853281] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61c41c3-4097-41eb-8c3a-cbc4033e0922 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.862810] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance '4edbf65d-eaca-47e1-b6bf-03ccb908b52e' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 809.892822] env[61995]: DEBUG nova.policy [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce97dbfe04a349738a0e5ee9cf177338', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a269227b55044d38154a50fbc722723', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 809.943766] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b7b695-76db-a278-08a4-d6ed1cd0737c/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 809.944452] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decaa2b1-17c8-41f6-bf39-528bd4fbf6f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.951380] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b7b695-76db-a278-08a4-d6ed1cd0737c/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 809.951633] env[61995]: ERROR oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b7b695-76db-a278-08a4-d6ed1cd0737c/disk-0.vmdk due to incomplete transfer. [ 809.951935] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5e006b8e-bea6-44ee-98db-79bcc96e1215 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.960440] env[61995]: DEBUG oslo_vmware.rw_handles [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b7b695-76db-a278-08a4-d6ed1cd0737c/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 809.960643] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Uploaded image 49cfe0c7-9c09-4a63-bd2f-322ba7cb789f to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 809.962644] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 809.965912] env[61995]: DEBUG oslo_concurrency.lockutils [req-30daf9ac-3ebd-46de-b3cb-c4a4a5b3f9ce req-a570f6bc-996f-4c11-90bd-b5d0ba12ab38 service nova] Releasing lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.965912] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4ec03985-7257-49df-9032-874a7b43b113 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.970417] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 809.970417] env[61995]: value = "task-794527" [ 809.970417] env[61995]: _type = "Task" [ 809.970417] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.979326] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794527, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.001436] env[61995]: DEBUG nova.network.neutron [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updated VIF entry in instance network info cache for port 52c15b2e-a283-4a39-9a38-d2ccd7da93a2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.002143] env[61995]: DEBUG nova.network.neutron [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [{"id": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "address": "fa:16:3e:0f:bb:93", "network": {"id": "751238cd-fe57-43f1-a0fe-f0282f02f685", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-258090236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7381568d0e44fbb8bb161e9d5e8bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52c15b2e-a2", "ovs_interfaceid": "52c15b2e-a283-4a39-9a38-d2ccd7da93a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.178429] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794523, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.262246] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794526, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.304105] env[61995]: DEBUG nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.376779] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 810.377990] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4816b9bb-4cf5-4d0f-a711-581d0c294b75 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.385836] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 810.385836] env[61995]: value = "task-794528" [ 810.385836] env[61995]: _type = "Task" [ 810.385836] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.401406] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794528, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.424074] env[61995]: DEBUG nova.network.neutron [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Successfully created port: 9ba91a20-0f20-4a6a-86c4-bbb63bf131fc {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.481548] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794527, 'name': Destroy_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.505220] env[61995]: DEBUG oslo_concurrency.lockutils [req-b72ce384-76ee-4ae7-8d6d-7a39ed4ffc57 req-3389a6a7-5719-4c37-a678-d5e610743c44 service nova] Releasing lock "refresh_cache-e53d10a3-1d16-498d-b4d8-abe7205410ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.578685] env[61995]: DEBUG nova.network.neutron [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updated VIF entry in instance network info cache for port ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.579140] env[61995]: DEBUG nova.network.neutron [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updating instance_info_cache with network_info: [{"id": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "address": "fa:16:3e:d6:67:26", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcee8fc-e0", "ovs_interfaceid": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.675900] env[61995]: DEBUG oslo_vmware.api [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794523, 'name': PowerOnVM_Task, 'duration_secs': 0.542724} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.676229] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.676420] env[61995]: INFO nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Took 8.78 seconds to spawn the instance on the hypervisor. [ 810.676596] env[61995]: DEBUG nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.677379] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1c4180-c725-4d9d-a1fa-18bc6bd2fc9d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.765913] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.766117] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.766174] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.766459] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.766823] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.766823] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.767064] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.767222] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.767395] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.767566] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.767746] env[61995]: DEBUG nova.virt.hardware [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.769916] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794526, 'name': CreateVM_Task, 'duration_secs': 0.559991} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.770780] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edbc422-58d4-4cb8-b2fb-da756adbadfa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.773635] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 810.774424] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.774638] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.774981] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.775967] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fecdf9e7-4fa1-4d8a-ac04-3b5e8bd76bb1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.787794] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fd7ce5-1281-4498-822b-ad794893e89a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.793066] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 810.793066] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529cf8cf-c189-ab1a-f3ed-86ae0745b749" [ 810.793066] env[61995]: _type = "Task" [ 810.793066] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.808383] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d348e89-4aa8-4594-b582-9fea690425bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.809030] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:3f:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3393eaad-1226-4e95-866a-c9c4ad8858c3', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.816480] env[61995]: DEBUG oslo.service.loopingcall [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.821155] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 810.824981] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-862953a4-fec3-4806-9582-e1d260f69558 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.840905] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529cf8cf-c189-ab1a-f3ed-86ae0745b749, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.847855] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920f4853-ca92-4bf4-a8d7-4f4c6bb9c260 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.852488] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.852488] env[61995]: value = "task-794529" [ 810.852488] env[61995]: _type = "Task" [ 810.852488] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.884680] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173be231-425b-49c9-8a58-31a67f3c6666 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.892632] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794529, 'name': CreateVM_Task} progress is 15%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.901782] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d6534f-48e7-4cfa-a2d2-48f7a53bcb8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.906208] env[61995]: DEBUG oslo_vmware.api [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794528, 'name': PowerOnVM_Task, 'duration_secs': 0.394852} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.906501] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.906691] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06d38739-4974-4b30-a739-f6d57d3ceb84 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance '4edbf65d-eaca-47e1-b6bf-03ccb908b52e' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 810.923447] env[61995]: DEBUG nova.compute.provider_tree [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.986643] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794527, 'name': Destroy_Task, 'duration_secs': 0.614926} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.986919] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Destroyed the VM [ 810.987272] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 810.987580] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e302514f-3ba7-41da-b1df-b0e4b1080464 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.993985] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 810.993985] env[61995]: value = "task-794530" [ 810.993985] env[61995]: _type = "Task" [ 810.993985] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.002985] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794530, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.085160] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f5eb224-c97b-4dce-9d67-b323184fe328 req-3a969c65-04f8-4da7-9a82-0ff46743a4a4 service nova] Releasing lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.199027] env[61995]: INFO nova.compute.manager [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Took 41.32 seconds to build instance. [ 811.303668] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529cf8cf-c189-ab1a-f3ed-86ae0745b749, 'name': SearchDatastore_Task, 'duration_secs': 0.057288} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.303939] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.304199] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.304441] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.304660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.304762] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.305028] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af9a937f-ab1b-4af7-9218-c037c39c1822 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.313531] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.313727] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 811.314452] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4960559-88b2-4336-a729-34265a1ff939 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.319455] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 811.319455] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5286deed-0c68-eeb4-9707-99a1ff0c2156" [ 811.319455] env[61995]: _type = "Task" [ 811.319455] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.323264] env[61995]: DEBUG nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.330420] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5286deed-0c68-eeb4-9707-99a1ff0c2156, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.352997] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.353271] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.353434] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.353622] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.353774] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.353926] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.354154] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.354324] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.354531] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.354754] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.354948] env[61995]: DEBUG nova.virt.hardware [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.355793] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91684bc4-f216-4a8c-bbae-1b7b6abf0a88 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.368601] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586df15c-941a-4e80-8a00-0cde8c899ef8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.371998] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794529, 'name': CreateVM_Task, 'duration_secs': 0.436137} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.372164] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 811.373098] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.373345] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.373684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.373909] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e2297ff-8cef-4818-929d-2bf4e738b414 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.385823] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 811.385823] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52308980-39c2-d8c3-2af1-5cc866316b9f" [ 811.385823] env[61995]: _type = "Task" [ 811.385823] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.393020] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52308980-39c2-d8c3-2af1-5cc866316b9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.426256] env[61995]: DEBUG nova.scheduler.client.report [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.504810] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794530, 'name': RemoveSnapshot_Task} progress is 47%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.636469] env[61995]: DEBUG nova.compute.manager [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Received event network-changed-b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.636829] env[61995]: DEBUG nova.compute.manager [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Refreshing instance network info cache due to event network-changed-b970a5e5-c432-4698-bc92-c64f8b7e3091. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 811.637223] env[61995]: DEBUG oslo_concurrency.lockutils [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] Acquiring lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.637517] env[61995]: DEBUG oslo_concurrency.lockutils [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] Acquired lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.637826] env[61995]: DEBUG nova.network.neutron [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Refreshing network info cache for port b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.701814] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8a9e173f-2544-4287-9167-eaf9dae2b04c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.972s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.831184] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5286deed-0c68-eeb4-9707-99a1ff0c2156, 'name': SearchDatastore_Task, 'duration_secs': 0.01754} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.831981] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f130e494-1679-481e-b1d2-d1ad1f25acce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.837214] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 811.837214] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ddf71b-549b-5b12-3a67-13a96d9bc6fa" [ 811.837214] env[61995]: _type = "Task" [ 811.837214] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.846248] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ddf71b-549b-5b12-3a67-13a96d9bc6fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.900832] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52308980-39c2-d8c3-2af1-5cc866316b9f, 'name': SearchDatastore_Task, 'duration_secs': 0.009135} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.901169] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.901420] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.902210] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.902378] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.902791] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.903130] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-565c965b-c160-4c92-a83d-918dcfab16da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.911614] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.911802] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 811.912518] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95917b9a-326d-44aa-b58b-8d66d4a0bfea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.926324] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 811.926324] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520bff19-ff52-65b6-9610-83967f549c4c" [ 811.926324] env[61995]: _type = "Task" [ 811.926324] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.934116] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.934640] env[61995]: DEBUG nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.941145] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.423s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.941442] env[61995]: DEBUG nova.objects.instance [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lazy-loading 'resources' on Instance uuid c51184b6-d2c7-49ff-b061-ab0f9a69cdbf {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.949227] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520bff19-ff52-65b6-9610-83967f549c4c, 'name': SearchDatastore_Task, 'duration_secs': 0.022978} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.949989] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8c7e7b9-c771-40b0-83da-811a0535934a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.959019] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 811.959019] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b58cbe-ca30-9678-899d-ab389d064622" [ 811.959019] env[61995]: _type = "Task" [ 811.959019] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.966920] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b58cbe-ca30-9678-899d-ab389d064622, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.006635] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794530, 'name': RemoveSnapshot_Task} progress is 47%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.130173] env[61995]: DEBUG nova.compute.manager [req-c7f0b043-5872-4504-ae42-8a8de8b9954c req-9223a226-88b0-42f4-9f66-3409cf038cd5 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Received event network-vif-plugged-9ba91a20-0f20-4a6a-86c4-bbb63bf131fc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.130733] env[61995]: DEBUG oslo_concurrency.lockutils [req-c7f0b043-5872-4504-ae42-8a8de8b9954c req-9223a226-88b0-42f4-9f66-3409cf038cd5 service nova] Acquiring lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.130733] env[61995]: DEBUG oslo_concurrency.lockutils [req-c7f0b043-5872-4504-ae42-8a8de8b9954c req-9223a226-88b0-42f4-9f66-3409cf038cd5 service nova] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.130818] env[61995]: DEBUG oslo_concurrency.lockutils [req-c7f0b043-5872-4504-ae42-8a8de8b9954c req-9223a226-88b0-42f4-9f66-3409cf038cd5 service nova] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.130928] env[61995]: DEBUG nova.compute.manager [req-c7f0b043-5872-4504-ae42-8a8de8b9954c req-9223a226-88b0-42f4-9f66-3409cf038cd5 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] No waiting events found dispatching network-vif-plugged-9ba91a20-0f20-4a6a-86c4-bbb63bf131fc {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.131240] env[61995]: WARNING nova.compute.manager [req-c7f0b043-5872-4504-ae42-8a8de8b9954c req-9223a226-88b0-42f4-9f66-3409cf038cd5 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Received unexpected event network-vif-plugged-9ba91a20-0f20-4a6a-86c4-bbb63bf131fc for instance with vm_state building and task_state spawning. [ 812.303368] env[61995]: DEBUG nova.network.neutron [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Successfully updated port: 9ba91a20-0f20-4a6a-86c4-bbb63bf131fc {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.352317] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ddf71b-549b-5b12-3a67-13a96d9bc6fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009454} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.352674] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.352943] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] b40b1866-ced3-40ef-9ab7-30d58b75f288/b40b1866-ced3-40ef-9ab7-30d58b75f288.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 812.353246] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c8c5921-25ed-4436-a3d5-f9bde1b457c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.362094] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 812.362094] env[61995]: value = "task-794531" [ 812.362094] env[61995]: _type = "Task" [ 812.362094] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.370525] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.406699] env[61995]: DEBUG nova.network.neutron [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Updated VIF entry in instance network info cache for port b970a5e5-c432-4698-bc92-c64f8b7e3091. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 812.407099] env[61995]: DEBUG nova.network.neutron [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Updating instance_info_cache with network_info: [{"id": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "address": "fa:16:3e:1e:e0:ed", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb970a5e5-c4", "ovs_interfaceid": "b970a5e5-c432-4698-bc92-c64f8b7e3091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.444804] env[61995]: DEBUG nova.compute.utils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.447132] env[61995]: DEBUG nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.447363] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 812.467132] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b58cbe-ca30-9678-899d-ab389d064622, 'name': SearchDatastore_Task, 'duration_secs': 0.009406} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.467432] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.467713] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 812.467959] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23e47e00-7880-41e4-adc8-20086f04b8b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.474371] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 812.474371] env[61995]: value = "task-794532" [ 812.474371] env[61995]: _type = "Task" [ 812.474371] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.485613] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794532, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.494590] env[61995]: DEBUG nova.policy [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9456ac4a7aa64ef09b256834206d2608', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f22192b1bd12430e9f2de0288d1e7e2a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 812.506956] env[61995]: DEBUG oslo_vmware.api [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794530, 'name': RemoveSnapshot_Task, 'duration_secs': 1.074419} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.507259] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 812.507535] env[61995]: INFO nova.compute.manager [None req-620e942c-94a4-4b60-9c5d-6980834de331 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Took 19.38 seconds to snapshot the instance on the hypervisor. [ 812.809869] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "refresh_cache-0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.810278] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "refresh_cache-0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.810447] env[61995]: DEBUG nova.network.neutron [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.822716] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Successfully created port: a502d606-f8c7-4230-a77e-4abdd17684e3 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.872724] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794531, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.874749] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd4384f-a742-43de-8eb5-6e8360e300c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.884219] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcaa713-c86e-477e-88a7-65feee23ecb8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.919311] env[61995]: DEBUG oslo_concurrency.lockutils [req-a4555c13-be90-4983-9b2b-8a42f9d2e46a req-af204538-88c9-46c7-ba54-555f4b1f8ad0 service nova] Releasing lock "refresh_cache-dc7077ac-d3fd-4e84-867a-d86328b32f5d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.920538] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a8afd2-4cda-4829-b8ce-510175175a92 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.930899] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73db367-d9fe-422c-9529-ebadb623717d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.946646] env[61995]: DEBUG nova.compute.provider_tree [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.953814] env[61995]: DEBUG nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.989045] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794532, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.210227] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.210227] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.210227] env[61995]: DEBUG nova.compute.manager [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Going to confirm migration 1 {{(pid=61995) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 813.215634] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Successfully created port: 8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.351733] env[61995]: DEBUG nova.network.neutron [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.381024] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5374} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.381024] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] b40b1866-ced3-40ef-9ab7-30d58b75f288/b40b1866-ced3-40ef-9ab7-30d58b75f288.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 813.381024] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.381024] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-695f58fc-7cf3-4b98-a21f-0dd288caf645 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.386174] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 813.386174] env[61995]: value = "task-794533" [ 813.386174] env[61995]: _type = "Task" [ 813.386174] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.394618] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794533, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.450582] env[61995]: DEBUG nova.scheduler.client.report [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.488882] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794532, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63947} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.490802] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 813.490802] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.490802] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03991649-2aa4-4d4e-bece-b5e9b79d4bb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.501057] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 813.501057] env[61995]: value = "task-794534" [ 813.501057] env[61995]: _type = "Task" [ 813.501057] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.510015] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794534, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.559919] env[61995]: DEBUG nova.network.neutron [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Updating instance_info_cache with network_info: [{"id": "9ba91a20-0f20-4a6a-86c4-bbb63bf131fc", "address": "fa:16:3e:df:58:75", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba91a20-0f", "ovs_interfaceid": "9ba91a20-0f20-4a6a-86c4-bbb63bf131fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.768044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.768297] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.768487] env[61995]: DEBUG nova.network.neutron [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 813.768675] env[61995]: DEBUG nova.objects.instance [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lazy-loading 'info_cache' on Instance uuid 4edbf65d-eaca-47e1-b6bf-03ccb908b52e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.895863] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794533, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065346} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.896170] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.896918] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a6d395-9fd5-4f44-8738-44a49b98b021 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.921089] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] b40b1866-ced3-40ef-9ab7-30d58b75f288/b40b1866-ced3-40ef-9ab7-30d58b75f288.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.922036] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d78c161e-5db3-47ee-950e-2a75b1fd1d21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.940455] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 813.940455] env[61995]: value = "task-794535" [ 813.940455] env[61995]: _type = "Task" [ 813.940455] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.948623] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794535, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.955969] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.016s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.958436] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.280s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.958724] env[61995]: DEBUG nova.objects.instance [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lazy-loading 'resources' on Instance uuid f6db353f-f7b5-4ba4-8f70-f3b65281da93 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.964181] env[61995]: DEBUG nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.981012] env[61995]: INFO nova.scheduler.client.report [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Deleted allocations for instance c51184b6-d2c7-49ff-b061-ab0f9a69cdbf [ 813.989225] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.989542] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.989673] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.990962] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.990962] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.990962] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.990962] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.990962] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.990962] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.990962] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.991295] env[61995]: DEBUG nova.virt.hardware [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.991919] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c8e058-96bd-474c-9af5-b441da9f5dd6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.004168] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13611291-a683-4bd7-8925-743f437987c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.014206] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794534, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.508272} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.023013] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.024243] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ba8486-48ec-42aa-861e-5c8431cbee2a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.046966] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.047937] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f6a5419-7714-49f4-a662-5de1c410bed2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.062156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "refresh_cache-0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.062469] env[61995]: DEBUG nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Instance network_info: |[{"id": "9ba91a20-0f20-4a6a-86c4-bbb63bf131fc", "address": "fa:16:3e:df:58:75", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba91a20-0f", "ovs_interfaceid": "9ba91a20-0f20-4a6a-86c4-bbb63bf131fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.062865] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:58:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ba91a20-0f20-4a6a-86c4-bbb63bf131fc', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.069961] env[61995]: DEBUG oslo.service.loopingcall [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.070777] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 814.070996] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4f3c971-c8c1-48f0-888e-eef620750311 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.085689] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 814.085689] env[61995]: value = "task-794536" [ 814.085689] env[61995]: _type = "Task" [ 814.085689] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.092779] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.092779] env[61995]: value = "task-794537" [ 814.092779] env[61995]: _type = "Task" [ 814.092779] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.096914] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.104531] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794537, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.129606] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "4da426a1-7af9-4e4e-82ed-c45687a19b86" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.129906] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.130160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "4da426a1-7af9-4e4e-82ed-c45687a19b86-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.130359] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.130559] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.135524] env[61995]: INFO nova.compute.manager [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Terminating instance [ 814.137367] env[61995]: DEBUG nova.compute.manager [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.137588] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 814.138449] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba34d5c5-34b3-4988-805b-9f65abf86896 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.145642] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 814.145868] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a3cadce-c4b2-4cc7-a2b1-8ca78186c945 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.151864] env[61995]: DEBUG oslo_vmware.api [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 814.151864] env[61995]: value = "task-794538" [ 814.151864] env[61995]: _type = "Task" [ 814.151864] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.158658] env[61995]: DEBUG nova.compute.manager [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Received event network-changed-9ba91a20-0f20-4a6a-86c4-bbb63bf131fc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.158907] env[61995]: DEBUG nova.compute.manager [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Refreshing instance network info cache due to event network-changed-9ba91a20-0f20-4a6a-86c4-bbb63bf131fc. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 814.159086] env[61995]: DEBUG oslo_concurrency.lockutils [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] Acquiring lock "refresh_cache-0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.159237] env[61995]: DEBUG oslo_concurrency.lockutils [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] Acquired lock "refresh_cache-0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.159401] env[61995]: DEBUG nova.network.neutron [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Refreshing network info cache for port 9ba91a20-0f20-4a6a-86c4-bbb63bf131fc {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.163771] env[61995]: DEBUG oslo_vmware.api [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794538, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.455602] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794535, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.490350] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d353977-40f3-465e-ab6a-a77070f65159 tempest-SecurityGroupsTestJSON-395600174 tempest-SecurityGroupsTestJSON-395600174-project-member] Lock "c51184b6-d2c7-49ff-b061-ab0f9a69cdbf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.658s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.598909] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.610214] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794537, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.666558] env[61995]: DEBUG oslo_vmware.api [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794538, 'name': PowerOffVM_Task, 'duration_secs': 0.284562} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.669340] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 814.669581] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 814.670411] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c9e2fde-ec77-4780-a904-6565061a7ec3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.758920] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 814.759502] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 814.759502] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Deleting the datastore file [datastore2] 4da426a1-7af9-4e4e-82ed-c45687a19b86 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.759844] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d63a201a-14b3-4779-8ff2-149e126ec7c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.772960] env[61995]: DEBUG oslo_vmware.api [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for the task: (returnval){ [ 814.772960] env[61995]: value = "task-794540" [ 814.772960] env[61995]: _type = "Task" [ 814.772960] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.787771] env[61995]: DEBUG oslo_vmware.api [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794540, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.841876] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0a3829-9e30-4cd9-ba43-2612fd923617 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.849487] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd19672a-298a-4927-a2cc-fce02a7f63bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.889783] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137f1b34-f6c5-4e76-9af8-d97b517e4f44 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.898748] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe17a397-ec86-4263-b6fd-3291077b088f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.916739] env[61995]: DEBUG nova.compute.provider_tree [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.955544] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794535, 'name': ReconfigVM_Task, 'duration_secs': 0.863585} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.958490] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Reconfigured VM instance instance-00000035 to attach disk [datastore2] b40b1866-ced3-40ef-9ab7-30d58b75f288/b40b1866-ced3-40ef-9ab7-30d58b75f288.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 814.959585] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02205abb-0b84-4491-b64d-7448dc8ecfe8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.966768] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 814.966768] env[61995]: value = "task-794541" [ 814.966768] env[61995]: _type = "Task" [ 814.966768] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.972335] env[61995]: DEBUG nova.compute.manager [req-4f29ecfd-7481-4065-a8d7-147d176874d8 req-b01a5db7-b8df-4669-b689-8ff10dcd0a52 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received event network-vif-plugged-a502d606-f8c7-4230-a77e-4abdd17684e3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.972573] env[61995]: DEBUG oslo_concurrency.lockutils [req-4f29ecfd-7481-4065-a8d7-147d176874d8 req-b01a5db7-b8df-4669-b689-8ff10dcd0a52 service nova] Acquiring lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.972692] env[61995]: DEBUG oslo_concurrency.lockutils [req-4f29ecfd-7481-4065-a8d7-147d176874d8 req-b01a5db7-b8df-4669-b689-8ff10dcd0a52 service nova] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.972910] env[61995]: DEBUG oslo_concurrency.lockutils [req-4f29ecfd-7481-4065-a8d7-147d176874d8 req-b01a5db7-b8df-4669-b689-8ff10dcd0a52 service nova] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.973119] env[61995]: DEBUG nova.compute.manager [req-4f29ecfd-7481-4065-a8d7-147d176874d8 req-b01a5db7-b8df-4669-b689-8ff10dcd0a52 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] No waiting events found dispatching network-vif-plugged-a502d606-f8c7-4230-a77e-4abdd17684e3 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.973297] env[61995]: WARNING nova.compute.manager [req-4f29ecfd-7481-4065-a8d7-147d176874d8 req-b01a5db7-b8df-4669-b689-8ff10dcd0a52 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received unexpected event network-vif-plugged-a502d606-f8c7-4230-a77e-4abdd17684e3 for instance with vm_state building and task_state spawning. [ 814.978878] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794541, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.979553] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Successfully updated port: a502d606-f8c7-4230-a77e-4abdd17684e3 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.098750] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794536, 'name': ReconfigVM_Task, 'duration_secs': 0.781712} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.099133] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84/823d2169-2bcb-4d5a-9677-cc4ac49a7e84.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.102557] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59a9e2d0-7e8b-40d2-9d13-0067adbcaf1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.109616] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794537, 'name': CreateVM_Task, 'duration_secs': 0.717796} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.111232] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 815.111232] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 815.111232] env[61995]: value = "task-794542" [ 815.111232] env[61995]: _type = "Task" [ 815.111232] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.111716] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.111883] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.112258] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.112518] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35e5f18f-197d-4706-a1a5-f6790d0278a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.119742] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 815.119742] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ed7b87-48d3-c603-49f1-892e702f593b" [ 815.119742] env[61995]: _type = "Task" [ 815.119742] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.122817] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794542, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.130664] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ed7b87-48d3-c603-49f1-892e702f593b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.141391] env[61995]: DEBUG nova.network.neutron [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Updated VIF entry in instance network info cache for port 9ba91a20-0f20-4a6a-86c4-bbb63bf131fc. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.141749] env[61995]: DEBUG nova.network.neutron [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Updating instance_info_cache with network_info: [{"id": "9ba91a20-0f20-4a6a-86c4-bbb63bf131fc", "address": "fa:16:3e:df:58:75", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba91a20-0f", "ovs_interfaceid": "9ba91a20-0f20-4a6a-86c4-bbb63bf131fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.231591] env[61995]: DEBUG nova.network.neutron [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance_info_cache with network_info: [{"id": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "address": "fa:16:3e:3d:94:75", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap591f35b8-ba", "ovs_interfaceid": "591f35b8-baf5-4da1-b479-dc99ae2f7b75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.282462] env[61995]: DEBUG oslo_vmware.api [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Task: {'id': task-794540, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266827} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.282841] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.283188] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.283387] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.283565] env[61995]: INFO nova.compute.manager [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Took 1.15 seconds to destroy the instance on the hypervisor. [ 815.283818] env[61995]: DEBUG oslo.service.loopingcall [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.284031] env[61995]: DEBUG nova.compute.manager [-] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.284133] env[61995]: DEBUG nova.network.neutron [-] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.420883] env[61995]: DEBUG nova.scheduler.client.report [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.480941] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794541, 'name': Rename_Task, 'duration_secs': 0.198333} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.480941] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 815.480941] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1ae37f8-c36d-46ef-9a05-bc13dcba04a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.489409] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 815.489409] env[61995]: value = "task-794543" [ 815.489409] env[61995]: _type = "Task" [ 815.489409] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.498133] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.621984] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794542, 'name': Rename_Task, 'duration_secs': 0.347096} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.622381] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 815.622656] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-050584ca-a4c5-45db-8181-38508d39e651 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.635631] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ed7b87-48d3-c603-49f1-892e702f593b, 'name': SearchDatastore_Task, 'duration_secs': 0.014003} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.636998] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.637258] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.637489] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.637657] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.637904] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.638307] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 815.638307] env[61995]: value = "task-794544" [ 815.638307] env[61995]: _type = "Task" [ 815.638307] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.638473] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62b4eb93-1a31-48c0-99e6-837df1754a6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.644860] env[61995]: DEBUG oslo_concurrency.lockutils [req-53f6940f-8cc9-48c7-a0fb-edd17360abb2 req-a545efd8-7471-43a9-892b-27360c9c1d11 service nova] Releasing lock "refresh_cache-0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.651210] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794544, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.652601] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.652802] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 815.653507] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4c03af0-ba3f-4e05-92a1-55997346bd22 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.659417] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 815.659417] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527b98b1-c77b-d88a-9f31-0168511e5416" [ 815.659417] env[61995]: _type = "Task" [ 815.659417] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.668140] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527b98b1-c77b-d88a-9f31-0168511e5416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.734254] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4edbf65d-eaca-47e1-b6bf-03ccb908b52e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.734552] env[61995]: DEBUG nova.objects.instance [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lazy-loading 'migration_context' on Instance uuid 4edbf65d-eaca-47e1-b6bf-03ccb908b52e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 815.929613] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.971s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.934201] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.377s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.936727] env[61995]: INFO nova.compute.claims [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.956955] env[61995]: INFO nova.scheduler.client.report [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Deleted allocations for instance f6db353f-f7b5-4ba4-8f70-f3b65281da93 [ 816.001891] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794543, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.151592] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794544, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.169849] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527b98b1-c77b-d88a-9f31-0168511e5416, 'name': SearchDatastore_Task, 'duration_secs': 0.012749} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.170821] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3414f9fc-dab1-4eb8-8c1a-d24306f8cf98 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.176405] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 816.176405] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c374a4-72f2-869a-c02d-e718981354c7" [ 816.176405] env[61995]: _type = "Task" [ 816.176405] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.185173] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c374a4-72f2-869a-c02d-e718981354c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.190796] env[61995]: DEBUG nova.compute.manager [req-293ea380-6d49-4b22-947b-a806d493420a req-9fb1ee19-6466-47e3-ada0-36bfedf8c175 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Received event network-vif-deleted-b040835e-914d-428e-9341-ee8c08259624 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.191009] env[61995]: INFO nova.compute.manager [req-293ea380-6d49-4b22-947b-a806d493420a req-9fb1ee19-6466-47e3-ada0-36bfedf8c175 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Neutron deleted interface b040835e-914d-428e-9341-ee8c08259624; detaching it from the instance and deleting it from the info cache [ 816.191228] env[61995]: DEBUG nova.network.neutron [req-293ea380-6d49-4b22-947b-a806d493420a req-9fb1ee19-6466-47e3-ada0-36bfedf8c175 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.215311] env[61995]: DEBUG nova.network.neutron [-] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.238900] env[61995]: DEBUG nova.objects.base [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Object Instance<4edbf65d-eaca-47e1-b6bf-03ccb908b52e> lazy-loaded attributes: info_cache,migration_context {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 816.240454] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd409c85-d0e9-4fac-ad31-1045be955a7b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.261457] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fb7687a-3820-4bf0-9b9c-0b939d00329e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.268381] env[61995]: DEBUG oslo_vmware.api [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 816.268381] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5285c50b-03ac-f6b6-bd72-e91138cd422f" [ 816.268381] env[61995]: _type = "Task" [ 816.268381] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.275970] env[61995]: DEBUG oslo_vmware.api [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5285c50b-03ac-f6b6-bd72-e91138cd422f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.465929] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94f32aba-07fc-4701-892e-09ddbde3595c tempest-InstanceActionsV221TestJSON-914204068 tempest-InstanceActionsV221TestJSON-914204068-project-member] Lock "f6db353f-f7b5-4ba4-8f70-f3b65281da93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.784s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.501338] env[61995]: DEBUG oslo_vmware.api [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794543, 'name': PowerOnVM_Task, 'duration_secs': 0.683725} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.501662] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 816.501962] env[61995]: INFO nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Took 10.33 seconds to spawn the instance on the hypervisor. [ 816.502186] env[61995]: DEBUG nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.502998] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496e76d3-c0b6-4b64-ad42-be4148bc4353 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.651350] env[61995]: DEBUG oslo_vmware.api [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794544, 'name': PowerOnVM_Task, 'duration_secs': 0.716694} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.651639] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 816.651854] env[61995]: DEBUG nova.compute.manager [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.652653] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc77abc9-c6c4-4964-ab4c-7182f05a8602 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.691065] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c374a4-72f2-869a-c02d-e718981354c7, 'name': SearchDatastore_Task, 'duration_secs': 0.011249} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.691065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.691065] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d/0c45e1fc-7b9a-4873-a381-ee2850ad5d4d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 816.691315] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a9ae896-b6cd-4a9b-a1de-d0582c5c534d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.693775] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5442906-d58d-402a-8491-f278379706f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.701229] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 816.701229] env[61995]: value = "task-794545" [ 816.701229] env[61995]: _type = "Task" [ 816.701229] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.707519] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd4df54-7c57-46aa-afb0-d36fe42aee1e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.719644] env[61995]: INFO nova.compute.manager [-] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Took 1.44 seconds to deallocate network for instance. [ 816.727938] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.741111] env[61995]: DEBUG nova.compute.manager [req-293ea380-6d49-4b22-947b-a806d493420a req-9fb1ee19-6466-47e3-ada0-36bfedf8c175 service nova] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Detach interface failed, port_id=b040835e-914d-428e-9341-ee8c08259624, reason: Instance 4da426a1-7af9-4e4e-82ed-c45687a19b86 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 816.779651] env[61995]: DEBUG oslo_vmware.api [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5285c50b-03ac-f6b6-bd72-e91138cd422f, 'name': SearchDatastore_Task, 'duration_secs': 0.010341} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.779916] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.008894] env[61995]: DEBUG nova.compute.manager [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received event network-changed-a502d606-f8c7-4230-a77e-4abdd17684e3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.009231] env[61995]: DEBUG nova.compute.manager [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Refreshing instance network info cache due to event network-changed-a502d606-f8c7-4230-a77e-4abdd17684e3. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 817.009532] env[61995]: DEBUG oslo_concurrency.lockutils [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] Acquiring lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.009705] env[61995]: DEBUG oslo_concurrency.lockutils [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] Acquired lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.009972] env[61995]: DEBUG nova.network.neutron [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Refreshing network info cache for port a502d606-f8c7-4230-a77e-4abdd17684e3 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 817.026277] env[61995]: INFO nova.compute.manager [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Took 42.94 seconds to build instance. [ 817.173690] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.213661] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794545, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.230183] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.384318] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe08c2ce-5826-4ab8-81ad-9f77c86377de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.392831] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ad621e-9fdb-4c87-b041-35272f75516d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.425010] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008092a4-03d0-475b-95af-e55ac874995d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.433987] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03bf7147-26ef-4028-97a4-61021e645c40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.448062] env[61995]: DEBUG nova.compute.provider_tree [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.524095] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Successfully updated port: 8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.528581] env[61995]: DEBUG oslo_concurrency.lockutils [None req-06633d88-d316-42c5-8a86-83e26c2ebeee tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.487s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.571390] env[61995]: DEBUG nova.network.neutron [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.704034] env[61995]: DEBUG nova.network.neutron [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.726934] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.758362} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.727891] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d/0c45e1fc-7b9a-4873-a381-ee2850ad5d4d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 817.730271] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.730881] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f1e9dbfa-2752-411e-830d-fc850c3092a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.738861] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 817.738861] env[61995]: value = "task-794546" [ 817.738861] env[61995]: _type = "Task" [ 817.738861] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.747469] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794546, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.951723] env[61995]: DEBUG nova.scheduler.client.report [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.031872] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.218618] env[61995]: DEBUG oslo_concurrency.lockutils [req-64e0d344-0a87-4405-ad8d-3ec9a1930f46 req-647153d7-3f97-4f6c-84b4-c31233d1db38 service nova] Releasing lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.218618] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.218618] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 818.239501] env[61995]: DEBUG nova.compute.manager [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received event network-vif-plugged-8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.239746] env[61995]: DEBUG oslo_concurrency.lockutils [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] Acquiring lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.240012] env[61995]: DEBUG oslo_concurrency.lockutils [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.242609] env[61995]: DEBUG oslo_concurrency.lockutils [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.242609] env[61995]: DEBUG nova.compute.manager [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] No waiting events found dispatching network-vif-plugged-8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 818.242771] env[61995]: WARNING nova.compute.manager [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received unexpected event network-vif-plugged-8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad for instance with vm_state building and task_state spawning. [ 818.242911] env[61995]: DEBUG nova.compute.manager [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received event network-changed-8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.244655] env[61995]: DEBUG nova.compute.manager [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Refreshing instance network info cache due to event network-changed-8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 818.244655] env[61995]: DEBUG oslo_concurrency.lockutils [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] Acquiring lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.254530] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794546, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079498} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.254910] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.256071] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75cb0f1-ed73-4a44-8d05-275ca3fae1f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.283224] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d/0c45e1fc-7b9a-4873-a381-ee2850ad5d4d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.283224] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1014421-b828-48af-89ec-47d2e1f44329 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.309194] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 818.309194] env[61995]: value = "task-794547" [ 818.309194] env[61995]: _type = "Task" [ 818.309194] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.320016] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794547, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.468189] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.469079] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.472898] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.533s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.474389] env[61995]: INFO nova.compute.claims [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.533495] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.534082] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.534082] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.534213] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.534337] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.537679] env[61995]: INFO nova.compute.manager [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Terminating instance [ 818.540243] env[61995]: DEBUG nova.compute.manager [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 818.540541] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 818.541475] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a9a282-b024-49ec-94be-fb140cb43993 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.550098] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 818.552926] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5824203c-9a0c-4dc8-acb3-816c3c128713 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.559019] env[61995]: DEBUG oslo_vmware.api [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 818.559019] env[61995]: value = "task-794548" [ 818.559019] env[61995]: _type = "Task" [ 818.559019] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.569815] env[61995]: DEBUG oslo_vmware.api [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.765829] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.826029] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794547, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.974509] env[61995]: DEBUG nova.compute.utils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.977096] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.977335] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 819.066834] env[61995]: DEBUG nova.policy [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1abff00389554af5af907b8a50fdd800', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cdb96d484bb47f99dfdeb5612095325', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 819.073770] env[61995]: DEBUG oslo_vmware.api [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794548, 'name': PowerOffVM_Task, 'duration_secs': 0.442967} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.074291] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 819.074386] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 819.075234] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10377d58-d4cd-4987-9de0-750b0c4c68e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.229127] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 819.229127] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 819.229127] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleting the datastore file [datastore1] 823d2169-2bcb-4d5a-9677-cc4ac49a7e84 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.229127] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22304398-96db-4a9f-bee3-bda6a434614f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.237564] env[61995]: DEBUG oslo_vmware.api [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 819.237564] env[61995]: value = "task-794550" [ 819.237564] env[61995]: _type = "Task" [ 819.237564] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.247298] env[61995]: DEBUG oslo_vmware.api [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.332024] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794547, 'name': ReconfigVM_Task, 'duration_secs': 0.648403} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.332024] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d/0c45e1fc-7b9a-4873-a381-ee2850ad5d4d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.332024] env[61995]: DEBUG nova.network.neutron [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Updating instance_info_cache with network_info: [{"id": "a502d606-f8c7-4230-a77e-4abdd17684e3", "address": "fa:16:3e:f9:7b:44", "network": {"id": "1e1870ee-be51-477c-9051-10e4214d21d0", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-649429003", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa502d606-f8", "ovs_interfaceid": "a502d606-f8c7-4230-a77e-4abdd17684e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "address": "fa:16:3e:a0:61:9d", "network": {"id": "96f9c86c-e29d-43ee-bdec-8911d44c41da", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-478568295", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e4a2146-6b", "ovs_interfaceid": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.333814] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b633b6a-b5c9-4157-b955-ffc980a7f8a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.341688] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 819.341688] env[61995]: value = "task-794551" [ 819.341688] env[61995]: _type = "Task" [ 819.341688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.352959] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794551, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.473992] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "2b40609a-0826-4a16-af28-c8926af6b646" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.474256] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.478040] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.524909] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Successfully created port: 2f7661f6-7c37-41c5-97d4-282602f6df05 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.753279] env[61995]: DEBUG oslo_vmware.api [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135336} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.753615] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.753862] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 819.754134] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 819.755465] env[61995]: INFO nova.compute.manager [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Took 1.21 seconds to destroy the instance on the hypervisor. [ 819.755465] env[61995]: DEBUG oslo.service.loopingcall [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.755465] env[61995]: DEBUG nova.compute.manager [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.755465] env[61995]: DEBUG nova.network.neutron [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 819.836352] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Releasing lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.837278] env[61995]: DEBUG nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Instance network_info: |[{"id": "a502d606-f8c7-4230-a77e-4abdd17684e3", "address": "fa:16:3e:f9:7b:44", "network": {"id": "1e1870ee-be51-477c-9051-10e4214d21d0", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-649429003", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa502d606-f8", "ovs_interfaceid": "a502d606-f8c7-4230-a77e-4abdd17684e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "address": "fa:16:3e:a0:61:9d", "network": {"id": "96f9c86c-e29d-43ee-bdec-8911d44c41da", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-478568295", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e4a2146-6b", "ovs_interfaceid": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 819.837278] env[61995]: DEBUG oslo_concurrency.lockutils [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] Acquired lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.837563] env[61995]: DEBUG nova.network.neutron [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Refreshing network info cache for port 8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.844738] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:7b:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ba3bd22-c936-470e-89bd-b3a5587e87a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a502d606-f8c7-4230-a77e-4abdd17684e3', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:61:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f762954-6ca5-4da5-bf0a-5d31c51ec570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.855694] env[61995]: DEBUG oslo.service.loopingcall [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.859754] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 819.866862] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2db8685a-07a2-4972-aa01-7fc90cb7b44c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.893633] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a975b622-b36b-4639-b652-86188998697e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.902499] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794551, 'name': Rename_Task, 'duration_secs': 0.20719} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.907094] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 819.907378] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.907378] env[61995]: value = "task-794552" [ 819.907378] env[61995]: _type = "Task" [ 819.907378] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.907576] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-373c333a-4a5a-4791-a25f-85bc937082c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.910109] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f78d053-c12c-4bb4-b74b-b38dfcbf49b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.950372] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794552, 'name': CreateVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.953579] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95edf224-c68d-4baa-99bc-753bfd6171d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.956479] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 819.956479] env[61995]: value = "task-794553" [ 819.956479] env[61995]: _type = "Task" [ 819.956479] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.963665] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d8aebd-c929-46b3-880c-04523a2ad533 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.971696] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794553, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.982090] env[61995]: DEBUG nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 819.988341] env[61995]: DEBUG nova.compute.provider_tree [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.321500] env[61995]: DEBUG nova.compute.manager [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Received event network-changed-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.324094] env[61995]: DEBUG nova.compute.manager [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Refreshing instance network info cache due to event network-changed-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.324094] env[61995]: DEBUG oslo_concurrency.lockutils [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] Acquiring lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.324094] env[61995]: DEBUG oslo_concurrency.lockutils [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] Acquired lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.324094] env[61995]: DEBUG nova.network.neutron [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Refreshing network info cache for port ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.435708] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794552, 'name': CreateVM_Task, 'duration_secs': 0.471133} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.435708] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 820.435708] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.435708] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.437415] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 820.437415] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a8af81c-68e6-4ac8-b032-740a79e34cdc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.441934] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 820.441934] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52828b46-143a-331b-0978-ff7682f978c1" [ 820.441934] env[61995]: _type = "Task" [ 820.441934] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.450536] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52828b46-143a-331b-0978-ff7682f978c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.465785] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794553, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.491629] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.498438] env[61995]: DEBUG nova.scheduler.client.report [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.541477] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.551049] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.551330] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.551494] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.551680] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.551831] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.551986] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.552229] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.552395] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.552573] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.552739] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.552917] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.553960] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf70432-ea55-4e32-80c8-70def047e1d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.563647] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f189ae7-ab7e-49c1-aebf-910f3a54caa4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.582572] env[61995]: DEBUG nova.network.neutron [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.707853] env[61995]: DEBUG nova.network.neutron [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Updated VIF entry in instance network info cache for port 8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.708444] env[61995]: DEBUG nova.network.neutron [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Updating instance_info_cache with network_info: [{"id": "a502d606-f8c7-4230-a77e-4abdd17684e3", "address": "fa:16:3e:f9:7b:44", "network": {"id": "1e1870ee-be51-477c-9051-10e4214d21d0", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-649429003", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa502d606-f8", "ovs_interfaceid": "a502d606-f8c7-4230-a77e-4abdd17684e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "address": "fa:16:3e:a0:61:9d", "network": {"id": "96f9c86c-e29d-43ee-bdec-8911d44c41da", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-478568295", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e4a2146-6b", "ovs_interfaceid": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.953394] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52828b46-143a-331b-0978-ff7682f978c1, 'name': SearchDatastore_Task, 'duration_secs': 0.01018} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.955800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.955800] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.955800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.955800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.955800] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.955800] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7689b4f-b186-4580-b8c8-31f33f68de6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.963412] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.963608] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 820.967371] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f86310b-7283-4d60-86f5-3324f1b2785b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.970224] env[61995]: DEBUG oslo_vmware.api [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794553, 'name': PowerOnVM_Task, 'duration_secs': 0.594589} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.972523] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 820.972783] env[61995]: INFO nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Took 9.65 seconds to spawn the instance on the hypervisor. [ 820.972920] env[61995]: DEBUG nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.974244] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8aecfb-a76f-47e7-8f00-672fd1f698cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.979254] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 820.979254] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529967ec-809e-065d-ecd0-1fbbdcc4ba96" [ 820.979254] env[61995]: _type = "Task" [ 820.979254] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.991876] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529967ec-809e-065d-ecd0-1fbbdcc4ba96, 'name': SearchDatastore_Task, 'duration_secs': 0.008702} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.992783] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a452456-fc1c-4f40-80a3-5b3572d07163 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.998312] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 820.998312] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52218e63-3291-0755-a09e-4aa66bd54ab8" [ 820.998312] env[61995]: _type = "Task" [ 820.998312] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.005628] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52218e63-3291-0755-a09e-4aa66bd54ab8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.010530] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.012103] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.015934] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.506s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.017797] env[61995]: INFO nova.compute.claims [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.084726] env[61995]: INFO nova.compute.manager [-] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Took 1.33 seconds to deallocate network for instance. [ 821.192579] env[61995]: DEBUG nova.network.neutron [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updated VIF entry in instance network info cache for port ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 821.192968] env[61995]: DEBUG nova.network.neutron [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updating instance_info_cache with network_info: [{"id": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "address": "fa:16:3e:d6:67:26", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcee8fc-e0", "ovs_interfaceid": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.211251] env[61995]: DEBUG oslo_concurrency.lockutils [req-10bb212f-90ba-4d06-9051-c1ff4595aa23 req-29b54dbb-dccc-4820-bae1-3fc682ad08bf service nova] Releasing lock "refresh_cache-0348098c-fd4c-41f8-a246-b9ba87a75674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.414039] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Successfully updated port: 2f7661f6-7c37-41c5-97d4-282602f6df05 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.494760] env[61995]: INFO nova.compute.manager [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Took 42.44 seconds to build instance. [ 821.508684] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52218e63-3291-0755-a09e-4aa66bd54ab8, 'name': SearchDatastore_Task, 'duration_secs': 0.009591} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.508952] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.509340] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 0348098c-fd4c-41f8-a246-b9ba87a75674/0348098c-fd4c-41f8-a246-b9ba87a75674.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 821.510253] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-edb7ad7e-cbf5-4c85-ae4a-6101135399e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.517302] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 821.517302] env[61995]: value = "task-794554" [ 821.517302] env[61995]: _type = "Task" [ 821.517302] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.522263] env[61995]: DEBUG nova.compute.utils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.525445] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 821.525615] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 821.538857] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.577100] env[61995]: DEBUG nova.policy [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1abff00389554af5af907b8a50fdd800', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cdb96d484bb47f99dfdeb5612095325', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 821.592514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.696734] env[61995]: DEBUG oslo_concurrency.lockutils [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] Releasing lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.699245] env[61995]: DEBUG nova.compute.manager [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Received event network-vif-deleted-3393eaad-1226-4e95-866a-c9c4ad8858c3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.699245] env[61995]: INFO nova.compute.manager [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Neutron deleted interface 3393eaad-1226-4e95-866a-c9c4ad8858c3; detaching it from the instance and deleting it from the info cache [ 821.699245] env[61995]: DEBUG nova.network.neutron [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.848314] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Successfully created port: 73460573-883d-4a1e-86f5-2c257ff21ea0 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.925260] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "refresh_cache-7524c912-4d73-462a-949f-3b4c85874094" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.925447] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "refresh_cache-7524c912-4d73-462a-949f-3b4c85874094" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.925735] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.997953] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb38d8f9-0d60-4ab6-8fe1-47257f667c21 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.030s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.035965] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.048993] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794554, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.204268] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0a1fb6ed-4b9f-4575-a50f-2755626189e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.219987] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53c5521-725b-4987-8cbb-7cb5db8666c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.261977] env[61995]: DEBUG nova.compute.manager [req-b66c7efc-4709-4af4-a21f-8956a0c0a6bc req-31d5488e-5e4c-45e0-b24f-830d4b843e6d service nova] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Detach interface failed, port_id=3393eaad-1226-4e95-866a-c9c4ad8858c3, reason: Instance 823d2169-2bcb-4d5a-9677-cc4ac49a7e84 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 822.475718] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.484685] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2111b1-2608-47bd-9f99-23441b56a649 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.502373] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46f8038-9177-4827-b16d-3e6fee41dce8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.541029] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be7d039-9c1f-45a7-a797-cad1afaa448d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.558672] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d11e58-af28-4f36-9eae-6d2765ba2643 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.563570] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794554, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589185} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.564141] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 0348098c-fd4c-41f8-a246-b9ba87a75674/0348098c-fd4c-41f8-a246-b9ba87a75674.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 822.564356] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.565037] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41a8d7c3-224b-4dbd-b701-4ab518b0585f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.576170] env[61995]: DEBUG nova.compute.provider_tree [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.586397] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 822.586397] env[61995]: value = "task-794555" [ 822.586397] env[61995]: _type = "Task" [ 822.586397] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.594265] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.636142] env[61995]: DEBUG nova.compute.manager [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Received event network-vif-plugged-2f7661f6-7c37-41c5-97d4-282602f6df05 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.636408] env[61995]: DEBUG oslo_concurrency.lockutils [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] Acquiring lock "7524c912-4d73-462a-949f-3b4c85874094-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.636485] env[61995]: DEBUG oslo_concurrency.lockutils [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] Lock "7524c912-4d73-462a-949f-3b4c85874094-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.636635] env[61995]: DEBUG oslo_concurrency.lockutils [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] Lock "7524c912-4d73-462a-949f-3b4c85874094-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.636816] env[61995]: DEBUG nova.compute.manager [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] No waiting events found dispatching network-vif-plugged-2f7661f6-7c37-41c5-97d4-282602f6df05 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.637101] env[61995]: WARNING nova.compute.manager [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Received unexpected event network-vif-plugged-2f7661f6-7c37-41c5-97d4-282602f6df05 for instance with vm_state building and task_state spawning. [ 822.637222] env[61995]: DEBUG nova.compute.manager [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Received event network-changed-2f7661f6-7c37-41c5-97d4-282602f6df05 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.637395] env[61995]: DEBUG nova.compute.manager [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Refreshing instance network info cache due to event network-changed-2f7661f6-7c37-41c5-97d4-282602f6df05. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 822.637577] env[61995]: DEBUG oslo_concurrency.lockutils [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] Acquiring lock "refresh_cache-7524c912-4d73-462a-949f-3b4c85874094" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.674634] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Updating instance_info_cache with network_info: [{"id": "2f7661f6-7c37-41c5-97d4-282602f6df05", "address": "fa:16:3e:5b:22:83", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f7661f6-7c", "ovs_interfaceid": "2f7661f6-7c37-41c5-97d4-282602f6df05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.068031] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.079917] env[61995]: DEBUG nova.scheduler.client.report [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.100651] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108105} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.106241] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.106413] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.106585] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.106873] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.106945] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.107137] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.107481] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.107537] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.107682] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.107863] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.108088] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.108415] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.109245] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777d2546-fd2e-420b-8510-d74de5403317 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.112903] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf59951-5159-4066-96aa-b96679164df6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.133725] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6285a53-e86b-44f9-9104-f26e47003b49 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.147090] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 0348098c-fd4c-41f8-a246-b9ba87a75674/0348098c-fd4c-41f8-a246-b9ba87a75674.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.147767] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eab7b96b-6865-4e4b-ad10-baf0e0314a56 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.176499] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 823.176499] env[61995]: value = "task-794556" [ 823.176499] env[61995]: _type = "Task" [ 823.176499] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.185562] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "refresh_cache-7524c912-4d73-462a-949f-3b4c85874094" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.185933] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Instance network_info: |[{"id": "2f7661f6-7c37-41c5-97d4-282602f6df05", "address": "fa:16:3e:5b:22:83", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f7661f6-7c", "ovs_interfaceid": "2f7661f6-7c37-41c5-97d4-282602f6df05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.186815] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794556, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.187095] env[61995]: DEBUG oslo_concurrency.lockutils [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] Acquired lock "refresh_cache-7524c912-4d73-462a-949f-3b4c85874094" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.187286] env[61995]: DEBUG nova.network.neutron [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Refreshing network info cache for port 2f7661f6-7c37-41c5-97d4-282602f6df05 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.188583] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:22:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0746f464-a938-427b-ba02-600449df5070', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f7661f6-7c37-41c5-97d4-282602f6df05', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.198560] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Creating folder: Project (6cdb96d484bb47f99dfdeb5612095325). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.201587] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30fe7ef5-bf61-443d-ad7a-81ceae825262 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.213815] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Created folder: Project (6cdb96d484bb47f99dfdeb5612095325) in parent group-v185203. [ 823.214186] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Creating folder: Instances. Parent ref: group-v185304. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.214356] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34bb3216-4ca8-42df-a423-554d4acf4ea7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.223694] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Created folder: Instances in parent group-v185304. [ 823.223977] env[61995]: DEBUG oslo.service.loopingcall [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.224215] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 823.224429] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e5b6dec-8e3b-45cd-af15-376b05cbf99d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.249390] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.249390] env[61995]: value = "task-794559" [ 823.249390] env[61995]: _type = "Task" [ 823.249390] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.259133] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794559, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.263658] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.263924] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.585966] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.586653] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.590376] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.877s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.590376] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.590644] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 823.595020] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.513s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.595020] env[61995]: INFO nova.compute.claims [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.596218] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82afe295-c127-4704-8a39-9c67e544ca32 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.605288] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3030f72c-3464-4d6b-85ea-2bc32d1e2fc3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.623185] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc51f07-dfc7-4ce2-8765-a842a332c4e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.632109] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb5ce27-333d-458d-8d79-4081c033fde0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.662144] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180290MB free_disk=66GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 823.662317] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.663425] env[61995]: DEBUG nova.network.neutron [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Updated VIF entry in instance network info cache for port 2f7661f6-7c37-41c5-97d4-282602f6df05. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.663758] env[61995]: DEBUG nova.network.neutron [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Updating instance_info_cache with network_info: [{"id": "2f7661f6-7c37-41c5-97d4-282602f6df05", "address": "fa:16:3e:5b:22:83", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f7661f6-7c", "ovs_interfaceid": "2f7661f6-7c37-41c5-97d4-282602f6df05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.687043] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794556, 'name': ReconfigVM_Task, 'duration_secs': 0.304135} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.687328] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 0348098c-fd4c-41f8-a246-b9ba87a75674/0348098c-fd4c-41f8-a246-b9ba87a75674.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.687947] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38615509-0abe-4748-9d29-057cad5c7d03 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.695770] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 823.695770] env[61995]: value = "task-794560" [ 823.695770] env[61995]: _type = "Task" [ 823.695770] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.707762] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794560, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.759953] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794559, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.765932] env[61995]: DEBUG nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.062988] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Successfully updated port: 73460573-883d-4a1e-86f5-2c257ff21ea0 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.099521] env[61995]: DEBUG nova.compute.utils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.104651] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.104706] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 824.167656] env[61995]: DEBUG oslo_concurrency.lockutils [req-75253f86-3d9c-4d7a-bf28-b276f89a0618 req-748e3b87-45c4-4c95-a602-a64af6e9fd4d service nova] Releasing lock "refresh_cache-7524c912-4d73-462a-949f-3b4c85874094" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.193110] env[61995]: DEBUG nova.policy [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1abff00389554af5af907b8a50fdd800', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cdb96d484bb47f99dfdeb5612095325', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 824.208057] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794560, 'name': Rename_Task, 'duration_secs': 0.156028} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.208390] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 824.208637] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb4e5071-13a0-4445-ba49-2b8189a07a1d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.215009] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 824.215009] env[61995]: value = "task-794561" [ 824.215009] env[61995]: _type = "Task" [ 824.215009] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.223216] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.259933] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794559, 'name': CreateVM_Task, 'duration_secs': 0.602506} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.260181] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 824.264022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.264022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.264022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.264022] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6d15d3b-00f4-4fe2-ba17-9d22827d95db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.266372] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 824.266372] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5295b40b-b9bd-e2b1-52b5-2466c43e9de6" [ 824.266372] env[61995]: _type = "Task" [ 824.266372] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.282090] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5295b40b-b9bd-e2b1-52b5-2466c43e9de6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.300575] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.566313] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "refresh_cache-bb31ef64-478f-4f39-9ee2-fcfabe8235ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.566477] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "refresh_cache-bb31ef64-478f-4f39-9ee2-fcfabe8235ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.566632] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 824.609639] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.611891] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Successfully created port: 6910a7c3-0d3f-4941-8135-018a19f8d5ef {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.664762] env[61995]: DEBUG nova.compute.manager [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Received event network-vif-plugged-73460573-883d-4a1e-86f5-2c257ff21ea0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.664993] env[61995]: DEBUG oslo_concurrency.lockutils [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] Acquiring lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.665453] env[61995]: DEBUG oslo_concurrency.lockutils [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.667031] env[61995]: DEBUG oslo_concurrency.lockutils [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.667031] env[61995]: DEBUG nova.compute.manager [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] No waiting events found dispatching network-vif-plugged-73460573-883d-4a1e-86f5-2c257ff21ea0 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.667031] env[61995]: WARNING nova.compute.manager [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Received unexpected event network-vif-plugged-73460573-883d-4a1e-86f5-2c257ff21ea0 for instance with vm_state building and task_state spawning. [ 824.667031] env[61995]: DEBUG nova.compute.manager [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Received event network-changed-73460573-883d-4a1e-86f5-2c257ff21ea0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.667031] env[61995]: DEBUG nova.compute.manager [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Refreshing instance network info cache due to event network-changed-73460573-883d-4a1e-86f5-2c257ff21ea0. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 824.667031] env[61995]: DEBUG oslo_concurrency.lockutils [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] Acquiring lock "refresh_cache-bb31ef64-478f-4f39-9ee2-fcfabe8235ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.737637] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794561, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.778158] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5295b40b-b9bd-e2b1-52b5-2466c43e9de6, 'name': SearchDatastore_Task, 'duration_secs': 0.025088} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.778492] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.778726] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.778972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.779156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.779345] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.779609] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9af505ab-28ca-4847-b06f-ffaf76ec66de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.789466] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.789757] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 824.790527] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2da3aa82-14fb-470f-b90e-4031662fc295 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.795690] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 824.795690] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a29d58-fa5c-6b4a-6b96-4bc59ef39e60" [ 824.795690] env[61995]: _type = "Task" [ 824.795690] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.804708] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a29d58-fa5c-6b4a-6b96-4bc59ef39e60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.047030] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4330fd-6b3d-4e23-a2a4-f51cfb56b171 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.055124] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd10887-fcb2-4dc8-babb-d22afb712e2c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.089773] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9659e5-e8e0-41eb-81a1-940813d23757 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.098041] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7ed5d7-8cfd-466f-87dd-de806e7baa06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.111756] env[61995]: DEBUG nova.compute.provider_tree [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.123704] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.229660] env[61995]: DEBUG oslo_vmware.api [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794561, 'name': PowerOnVM_Task, 'duration_secs': 0.774799} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.229938] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 825.234022] env[61995]: INFO nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Took 11.27 seconds to spawn the instance on the hypervisor. [ 825.234022] env[61995]: DEBUG nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.234022] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b4b6a3-4fb9-4a51-b564-0a613c222984 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.276438] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Updating instance_info_cache with network_info: [{"id": "73460573-883d-4a1e-86f5-2c257ff21ea0", "address": "fa:16:3e:4f:72:d4", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73460573-88", "ovs_interfaceid": "73460573-883d-4a1e-86f5-2c257ff21ea0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.294495] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "229d966c-b70d-4a51-a176-2e88488d5c4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.294727] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.306203] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a29d58-fa5c-6b4a-6b96-4bc59ef39e60, 'name': SearchDatastore_Task, 'duration_secs': 0.008961} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.307015] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1252f03-6457-43cd-b3e8-563b10c952e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.312863] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 825.312863] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521148ef-1c1b-27dc-d419-80cf2b92638f" [ 825.312863] env[61995]: _type = "Task" [ 825.312863] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.321692] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521148ef-1c1b-27dc-d419-80cf2b92638f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.614449] env[61995]: DEBUG nova.scheduler.client.report [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.620087] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.644090] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.644359] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.644523] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.644712] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.644865] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.645057] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.645295] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.645465] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.645637] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.645805] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.645982] env[61995]: DEBUG nova.virt.hardware [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.646881] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108967ce-9f6c-456b-a9da-e18d3f3fca95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.655096] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a9a8b9-49eb-4e69-98f2-4ba9260db104 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.747836] env[61995]: INFO nova.compute.manager [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Took 39.65 seconds to build instance. [ 825.779177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "refresh_cache-bb31ef64-478f-4f39-9ee2-fcfabe8235ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.779610] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Instance network_info: |[{"id": "73460573-883d-4a1e-86f5-2c257ff21ea0", "address": "fa:16:3e:4f:72:d4", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73460573-88", "ovs_interfaceid": "73460573-883d-4a1e-86f5-2c257ff21ea0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.779934] env[61995]: DEBUG oslo_concurrency.lockutils [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] Acquired lock "refresh_cache-bb31ef64-478f-4f39-9ee2-fcfabe8235ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.780143] env[61995]: DEBUG nova.network.neutron [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Refreshing network info cache for port 73460573-883d-4a1e-86f5-2c257ff21ea0 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 825.781245] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:72:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0746f464-a938-427b-ba02-600449df5070', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73460573-883d-4a1e-86f5-2c257ff21ea0', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.788493] env[61995]: DEBUG oslo.service.loopingcall [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.791238] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 825.793594] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2e02370-2656-4e94-8a2b-ef419a7c95c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.807638] env[61995]: DEBUG nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.818922] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.818922] env[61995]: value = "task-794562" [ 825.818922] env[61995]: _type = "Task" [ 825.818922] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.826244] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521148ef-1c1b-27dc-d419-80cf2b92638f, 'name': SearchDatastore_Task, 'duration_secs': 0.023139} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.826879] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.827277] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 7524c912-4d73-462a-949f-3b4c85874094/7524c912-4d73-462a-949f-3b4c85874094.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 825.827467] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49161e33-1c4d-40f5-bb97-33b6260a0b33 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.832563] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794562, 'name': CreateVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.837557] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 825.837557] env[61995]: value = "task-794563" [ 825.837557] env[61995]: _type = "Task" [ 825.837557] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.846331] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.121282] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.122628] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.528s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.128487] env[61995]: INFO nova.compute.claims [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.200044] env[61995]: DEBUG nova.network.neutron [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Updated VIF entry in instance network info cache for port 73460573-883d-4a1e-86f5-2c257ff21ea0. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 826.200455] env[61995]: DEBUG nova.network.neutron [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Updating instance_info_cache with network_info: [{"id": "73460573-883d-4a1e-86f5-2c257ff21ea0", "address": "fa:16:3e:4f:72:d4", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73460573-88", "ovs_interfaceid": "73460573-883d-4a1e-86f5-2c257ff21ea0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.250631] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cd58da8c-61b6-4a61-98b3-84ccd4526f02 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.077s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.335627] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794562, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.343368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.350220] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794563, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.509896] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Successfully updated port: 6910a7c3-0d3f-4941-8135-018a19f8d5ef {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.639863] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "87709bb0-1651-461b-be6b-b64bcbd35901" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.640180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "87709bb0-1651-461b-be6b-b64bcbd35901" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.688587] env[61995]: DEBUG nova.compute.manager [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Received event network-vif-plugged-6910a7c3-0d3f-4941-8135-018a19f8d5ef {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.688828] env[61995]: DEBUG oslo_concurrency.lockutils [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] Acquiring lock "af7163e8-4e77-4336-80e9-0623e29ed8c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.689024] env[61995]: DEBUG oslo_concurrency.lockutils [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.689227] env[61995]: DEBUG oslo_concurrency.lockutils [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.689367] env[61995]: DEBUG nova.compute.manager [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] No waiting events found dispatching network-vif-plugged-6910a7c3-0d3f-4941-8135-018a19f8d5ef {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 826.689539] env[61995]: WARNING nova.compute.manager [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Received unexpected event network-vif-plugged-6910a7c3-0d3f-4941-8135-018a19f8d5ef for instance with vm_state building and task_state spawning. [ 826.689703] env[61995]: DEBUG nova.compute.manager [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Received event network-changed-6910a7c3-0d3f-4941-8135-018a19f8d5ef {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.689861] env[61995]: DEBUG nova.compute.manager [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Refreshing instance network info cache due to event network-changed-6910a7c3-0d3f-4941-8135-018a19f8d5ef. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 826.690066] env[61995]: DEBUG oslo_concurrency.lockutils [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] Acquiring lock "refresh_cache-af7163e8-4e77-4336-80e9-0623e29ed8c0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.690215] env[61995]: DEBUG oslo_concurrency.lockutils [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] Acquired lock "refresh_cache-af7163e8-4e77-4336-80e9-0623e29ed8c0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.690375] env[61995]: DEBUG nova.network.neutron [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Refreshing network info cache for port 6910a7c3-0d3f-4941-8135-018a19f8d5ef {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 826.705097] env[61995]: DEBUG oslo_concurrency.lockutils [req-fc39649b-5f97-4880-a8c8-3a6d25e513aa req-f6bb2564-9471-45b1-b049-60e2defd6279 service nova] Releasing lock "refresh_cache-bb31ef64-478f-4f39-9ee2-fcfabe8235ec" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.836841] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794562, 'name': CreateVM_Task, 'duration_secs': 0.689108} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.836841] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 826.836841] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.836841] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.836841] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.836841] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c82a85ae-02ab-4b03-8be2-8e52f3bc7e76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.843319] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 826.843319] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520fad90-f367-7724-9e79-a2c4b2d29319" [ 826.843319] env[61995]: _type = "Task" [ 826.843319] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.849545] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794563, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619411} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.850107] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 7524c912-4d73-462a-949f-3b4c85874094/7524c912-4d73-462a-949f-3b4c85874094.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 826.850329] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.850563] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c71e4c0-b97b-4843-9c51-f88efe2185e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.854921] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520fad90-f367-7724-9e79-a2c4b2d29319, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.861966] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 826.861966] env[61995]: value = "task-794564" [ 826.861966] env[61995]: _type = "Task" [ 826.861966] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.869099] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794564, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.013756] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "refresh_cache-af7163e8-4e77-4336-80e9-0623e29ed8c0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.025590] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "0348098c-fd4c-41f8-a246-b9ba87a75674" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.025898] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.026172] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.026556] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.026785] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.029092] env[61995]: INFO nova.compute.manager [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Terminating instance [ 827.031354] env[61995]: DEBUG nova.compute.manager [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.031562] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 827.032402] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e2f9d4-7b41-4e63-a02f-e8ccbdca348e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.040550] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 827.040803] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5f8b4fc-7fef-4e4c-95b9-91be13ff0d18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.047044] env[61995]: DEBUG oslo_vmware.api [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 827.047044] env[61995]: value = "task-794565" [ 827.047044] env[61995]: _type = "Task" [ 827.047044] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.055412] env[61995]: DEBUG oslo_vmware.api [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.144545] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "87709bb0-1651-461b-be6b-b64bcbd35901" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.145208] env[61995]: DEBUG nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.227018] env[61995]: DEBUG nova.network.neutron [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.336024] env[61995]: DEBUG nova.network.neutron [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.354784] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520fad90-f367-7724-9e79-a2c4b2d29319, 'name': SearchDatastore_Task, 'duration_secs': 0.027273} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.357258] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.357497] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.357734] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.357889] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.358119] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.358846] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f478adf-86ef-47e1-9d78-baf2f2a200f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.368165] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.368351] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 827.374640] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb6430b5-e25b-4c80-9d59-b0c1c56ac2dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.377092] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794564, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073669} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.378116] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.378633] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b83663-12c8-4565-9fac-02905650a8ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.382192] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 827.382192] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52db2985-6a39-210b-17d7-6f0817f0dd58" [ 827.382192] env[61995]: _type = "Task" [ 827.382192] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.402672] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 7524c912-4d73-462a-949f-3b4c85874094/7524c912-4d73-462a-949f-3b4c85874094.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.405436] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0a9038d-3988-4063-b118-34e10c768da3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.423086] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52db2985-6a39-210b-17d7-6f0817f0dd58, 'name': SearchDatastore_Task, 'duration_secs': 0.008566} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.424346] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1edd4c5-30d1-410c-93c5-5cc58038a70d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.428185] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 827.428185] env[61995]: value = "task-794566" [ 827.428185] env[61995]: _type = "Task" [ 827.428185] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.432704] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 827.432704] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52622397-b70f-91f9-03b5-3ffd93b96c8f" [ 827.432704] env[61995]: _type = "Task" [ 827.432704] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.443728] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794566, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.451404] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52622397-b70f-91f9-03b5-3ffd93b96c8f, 'name': SearchDatastore_Task, 'duration_secs': 0.009506} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.451670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.451938] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] bb31ef64-478f-4f39-9ee2-fcfabe8235ec/bb31ef64-478f-4f39-9ee2-fcfabe8235ec.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.452214] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ec2b503-3fb0-4af1-ba32-4623e38249ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.458422] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 827.458422] env[61995]: value = "task-794567" [ 827.458422] env[61995]: _type = "Task" [ 827.458422] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.470311] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.558574] env[61995]: DEBUG oslo_vmware.api [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794565, 'name': PowerOffVM_Task, 'duration_secs': 0.197108} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.559110] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 827.559305] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 827.559585] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-096d97b3-f658-4f9b-8d6e-85149d7a530f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.584661] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109e24d5-3a80-494b-9505-33be49aa14c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.594101] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23326084-eb27-4265-8c45-b9dbf1bc5c31 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.627835] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ada1751-db6b-4723-8e40-6671dbc39cdf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.636330] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f9f0fb-e4dc-4048-b39c-780f502da5cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.655419] env[61995]: DEBUG nova.compute.utils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.657502] env[61995]: DEBUG nova.compute.provider_tree [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.658926] env[61995]: DEBUG nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.659086] env[61995]: DEBUG nova.network.neutron [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 827.738517] env[61995]: DEBUG nova.policy [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0eff1b3a788846aaa39101f7e81c3ca7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fb7e9f4ff4a4af8a216b3e38e864561', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 827.780446] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 827.780679] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 827.780924] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Deleting the datastore file [datastore2] 0348098c-fd4c-41f8-a246-b9ba87a75674 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.781204] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96637460-ff5b-477f-b5e9-e73ffceeafa7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.789747] env[61995]: DEBUG oslo_vmware.api [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for the task: (returnval){ [ 827.789747] env[61995]: value = "task-794569" [ 827.789747] env[61995]: _type = "Task" [ 827.789747] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.798591] env[61995]: DEBUG oslo_vmware.api [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.842368] env[61995]: DEBUG oslo_concurrency.lockutils [req-0aec06ec-b5ed-444d-8aba-460bf9172ec3 req-acd2a6b8-5aca-4fee-b264-f3bc05566c50 service nova] Releasing lock "refresh_cache-af7163e8-4e77-4336-80e9-0623e29ed8c0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.842788] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "refresh_cache-af7163e8-4e77-4336-80e9-0623e29ed8c0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.842950] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 827.940420] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.968524] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794567, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.105650] env[61995]: DEBUG nova.network.neutron [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Successfully created port: 67d6315d-2ed0-45b1-b08e-d4967d01eae3 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.160121] env[61995]: DEBUG nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.163744] env[61995]: DEBUG nova.scheduler.client.report [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.299936] env[61995]: DEBUG oslo_vmware.api [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Task: {'id': task-794569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.286321} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.299936] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.300177] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 828.300435] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 828.300638] env[61995]: INFO nova.compute.manager [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Took 1.27 seconds to destroy the instance on the hypervisor. [ 828.301760] env[61995]: DEBUG oslo.service.loopingcall [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.301760] env[61995]: DEBUG nova.compute.manager [-] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.301760] env[61995]: DEBUG nova.network.neutron [-] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 828.404138] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.438861] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794566, 'name': ReconfigVM_Task, 'duration_secs': 0.653883} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.439244] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 7524c912-4d73-462a-949f-3b4c85874094/7524c912-4d73-462a-949f-3b4c85874094.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.439885] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25e23424-0023-4329-a8c6-c275843c6fd6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.449503] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 828.449503] env[61995]: value = "task-794570" [ 828.449503] env[61995]: _type = "Task" [ 828.449503] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.458796] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794570, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.470515] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512221} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.470914] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] bb31ef64-478f-4f39-9ee2-fcfabe8235ec/bb31ef64-478f-4f39-9ee2-fcfabe8235ec.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.471747] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.471747] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c06ecc90-31d4-4943-aee3-87e854f06352 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.480262] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 828.480262] env[61995]: value = "task-794571" [ 828.480262] env[61995]: _type = "Task" [ 828.480262] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.491273] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.627014] env[61995]: DEBUG nova.network.neutron [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Updating instance_info_cache with network_info: [{"id": "6910a7c3-0d3f-4941-8135-018a19f8d5ef", "address": "fa:16:3e:89:52:f1", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6910a7c3-0d", "ovs_interfaceid": "6910a7c3-0d3f-4941-8135-018a19f8d5ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.672223] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.672555] env[61995]: DEBUG nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.675036] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.387s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.675317] env[61995]: DEBUG nova.objects.instance [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lazy-loading 'resources' on Instance uuid d056dadd-492d-4a4d-abc2-fc5f113f865b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 828.809842] env[61995]: DEBUG nova.compute.manager [req-6e81dd41-a726-456b-a08e-3052ef04578b req-96bfd5be-c309-40ca-8a2f-eb8357dc4324 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received event network-vif-deleted-a502d606-f8c7-4230-a77e-4abdd17684e3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.809960] env[61995]: INFO nova.compute.manager [req-6e81dd41-a726-456b-a08e-3052ef04578b req-96bfd5be-c309-40ca-8a2f-eb8357dc4324 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Neutron deleted interface a502d606-f8c7-4230-a77e-4abdd17684e3; detaching it from the instance and deleting it from the info cache [ 828.810256] env[61995]: DEBUG nova.network.neutron [req-6e81dd41-a726-456b-a08e-3052ef04578b req-96bfd5be-c309-40ca-8a2f-eb8357dc4324 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Updating instance_info_cache with network_info: [{"id": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "address": "fa:16:3e:a0:61:9d", "network": {"id": "96f9c86c-e29d-43ee-bdec-8911d44c41da", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-478568295", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f22192b1bd12430e9f2de0288d1e7e2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e4a2146-6b", "ovs_interfaceid": "8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.960121] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794570, 'name': Rename_Task, 'duration_secs': 0.359894} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.960426] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 828.960694] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-208954f4-17b5-4f84-8513-8b5f5f10440f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.967273] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 828.967273] env[61995]: value = "task-794572" [ 828.967273] env[61995]: _type = "Task" [ 828.967273] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.976692] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.992184] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064901} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.992459] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.993224] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5010363e-1006-4294-bfc9-c46c87712983 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.015413] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] bb31ef64-478f-4f39-9ee2-fcfabe8235ec/bb31ef64-478f-4f39-9ee2-fcfabe8235ec.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.015729] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47360e1c-387f-4901-87c8-c577e2269db7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.038484] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 829.038484] env[61995]: value = "task-794573" [ 829.038484] env[61995]: _type = "Task" [ 829.038484] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.046809] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.131806] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "refresh_cache-af7163e8-4e77-4336-80e9-0623e29ed8c0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.132204] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Instance network_info: |[{"id": "6910a7c3-0d3f-4941-8135-018a19f8d5ef", "address": "fa:16:3e:89:52:f1", "network": {"id": "779c3af9-442e-46de-be48-c63429bbf452", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1952022978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdb96d484bb47f99dfdeb5612095325", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6910a7c3-0d", "ovs_interfaceid": "6910a7c3-0d3f-4941-8135-018a19f8d5ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.132684] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:52:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0746f464-a938-427b-ba02-600449df5070', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6910a7c3-0d3f-4941-8135-018a19f8d5ef', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.140521] env[61995]: DEBUG oslo.service.loopingcall [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.140759] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 829.140992] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91b61d55-b6ec-4fe5-94ad-f84c366d5a43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.161534] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.161534] env[61995]: value = "task-794574" [ 829.161534] env[61995]: _type = "Task" [ 829.161534] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.169343] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794574, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.176726] env[61995]: DEBUG nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.179665] env[61995]: DEBUG nova.compute.utils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.183285] env[61995]: DEBUG nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Not allocating networking since 'none' was specified. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 829.206872] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.207179] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.207371] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.207569] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.207781] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.208023] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.208246] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.208546] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.208678] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.208881] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.209130] env[61995]: DEBUG nova.virt.hardware [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.210066] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9ec923-087a-410f-87b0-40daf7927958 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.221120] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab05796e-41f7-45ea-b051-977586df458c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.314876] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b685159-3400-42ac-bca2-a73f9cebf50e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.324677] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a68d88-bace-437b-a26a-e17cad17f7d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.361551] env[61995]: DEBUG nova.compute.manager [req-6e81dd41-a726-456b-a08e-3052ef04578b req-96bfd5be-c309-40ca-8a2f-eb8357dc4324 service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Detach interface failed, port_id=a502d606-f8c7-4230-a77e-4abdd17684e3, reason: Instance 0348098c-fd4c-41f8-a246-b9ba87a75674 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 829.443103] env[61995]: DEBUG nova.network.neutron [-] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.477575] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794572, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.548103] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794573, 'name': ReconfigVM_Task, 'duration_secs': 0.292486} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.550810] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Reconfigured VM instance instance-00000039 to attach disk [datastore1] bb31ef64-478f-4f39-9ee2-fcfabe8235ec/bb31ef64-478f-4f39-9ee2-fcfabe8235ec.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.551685] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a80d271-5e0d-4bae-acb0-c9ad0cf2aa77 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.558065] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 829.558065] env[61995]: value = "task-794575" [ 829.558065] env[61995]: _type = "Task" [ 829.558065] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.568025] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794575, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.617661] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341fa37e-bd0d-4bb1-b313-ba8c5647c714 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.627599] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8362e9e-abe2-4588-8ea0-99e30fbef553 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.660338] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87cc5a2-9c89-4027-8c63-59496286d48b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.674222] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e45ba05-bde8-44a6-b10a-2a360410f597 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.678470] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794574, 'name': CreateVM_Task, 'duration_secs': 0.350235} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.678733] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 829.679728] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.679924] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.680360] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.680667] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2804b688-d4e6-4b58-88bb-2e719c234df7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.695809] env[61995]: DEBUG nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.696153] env[61995]: DEBUG nova.compute.provider_tree [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.703715] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 829.703715] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f14547-d0ed-6331-5e66-a13d2d4ba500" [ 829.703715] env[61995]: _type = "Task" [ 829.703715] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.714930] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f14547-d0ed-6331-5e66-a13d2d4ba500, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.840323] env[61995]: DEBUG nova.network.neutron [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Successfully updated port: 67d6315d-2ed0-45b1-b08e-d4967d01eae3 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.947578] env[61995]: INFO nova.compute.manager [-] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Took 1.65 seconds to deallocate network for instance. [ 829.978561] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794572, 'name': PowerOnVM_Task, 'duration_secs': 0.557811} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.978853] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 829.979167] env[61995]: INFO nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Took 9.49 seconds to spawn the instance on the hypervisor. [ 829.979377] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.980698] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d2374d-52f4-412b-a7f0-869e13967bcb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.068011] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794575, 'name': Rename_Task, 'duration_secs': 0.1468} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.068573] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.068831] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e91f2b5-c2d6-4842-a59c-be06aca16344 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.074444] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 830.074444] env[61995]: value = "task-794576" [ 830.074444] env[61995]: _type = "Task" [ 830.074444] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.081658] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794576, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.205072] env[61995]: DEBUG nova.scheduler.client.report [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.218091] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f14547-d0ed-6331-5e66-a13d2d4ba500, 'name': SearchDatastore_Task, 'duration_secs': 0.020436} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.218918] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.218918] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.219435] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.219435] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.219435] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.220203] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d252f88-ef11-4d8b-9f10-983a486f33d1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.228583] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.228713] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.229695] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd5b1ee9-3edf-4c91-a379-bf9ec75bede2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.235131] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 830.235131] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a9753b-694d-baa3-491b-daf5479d3063" [ 830.235131] env[61995]: _type = "Task" [ 830.235131] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.243810] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a9753b-694d-baa3-491b-daf5479d3063, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.345090] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "refresh_cache-451e226c-5fb7-4042-a706-3e58b0606546" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.345263] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquired lock "refresh_cache-451e226c-5fb7-4042-a706-3e58b0606546" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.345422] env[61995]: DEBUG nova.network.neutron [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.455348] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.498934] env[61995]: INFO nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Took 39.97 seconds to build instance. [ 830.585274] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794576, 'name': PowerOnVM_Task, 'duration_secs': 0.475629} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.585541] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.585746] env[61995]: INFO nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Took 7.52 seconds to spawn the instance on the hypervisor. [ 830.585962] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.586712] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2863d128-7f35-4407-9637-e56f32b4d4c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.709948] env[61995]: DEBUG nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.713266] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.038s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.715456] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.718s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.717097] env[61995]: INFO nova.compute.claims [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.736733] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.737321] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.737321] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.737526] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.737526] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.737637] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.737852] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.738966] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.738966] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.738966] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.738966] env[61995]: DEBUG nova.virt.hardware [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.739586] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e844dfa5-9e83-48a4-8d72-c4188c171c20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.742861] env[61995]: INFO nova.scheduler.client.report [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Deleted allocations for instance d056dadd-492d-4a4d-abc2-fc5f113f865b [ 830.758537] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a117ac-482c-43f5-85e4-097054f2bb69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.763524] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a9753b-694d-baa3-491b-daf5479d3063, 'name': SearchDatastore_Task, 'duration_secs': 0.015659} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.764778] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e772b693-01d6-42ac-90d0-bb94dbb0ef28 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.775520] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.780880] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Creating folder: Project (59a0d86b65dc46ff924202140805a226). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 830.782466] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05d7e4a4-565f-4f51-a42f-8be3af215e85 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.785376] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 830.785376] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f7780b-635a-641d-599a-927e33af1354" [ 830.785376] env[61995]: _type = "Task" [ 830.785376] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.794253] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f7780b-635a-641d-599a-927e33af1354, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.795524] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Created folder: Project (59a0d86b65dc46ff924202140805a226) in parent group-v185203. [ 830.795709] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Creating folder: Instances. Parent ref: group-v185309. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 830.795958] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-875d374f-18fb-4d28-bcb6-7e3bc1083ff2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.806043] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Created folder: Instances in parent group-v185309. [ 830.806043] env[61995]: DEBUG oslo.service.loopingcall [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.806043] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.806043] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd72646e-75dc-4a75-bb7c-e79c19689e1d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.823442] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.823442] env[61995]: value = "task-794579" [ 830.823442] env[61995]: _type = "Task" [ 830.823442] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.831716] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794579, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.845570] env[61995]: DEBUG nova.compute.manager [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Received event network-vif-deleted-8e4a2146-6bfc-4f6d-9f06-80b9bddf2dad {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.845763] env[61995]: DEBUG nova.compute.manager [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Received event network-vif-plugged-67d6315d-2ed0-45b1-b08e-d4967d01eae3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.849034] env[61995]: DEBUG oslo_concurrency.lockutils [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] Acquiring lock "451e226c-5fb7-4042-a706-3e58b0606546-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.849034] env[61995]: DEBUG oslo_concurrency.lockutils [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] Lock "451e226c-5fb7-4042-a706-3e58b0606546-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.849034] env[61995]: DEBUG oslo_concurrency.lockutils [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] Lock "451e226c-5fb7-4042-a706-3e58b0606546-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.849034] env[61995]: DEBUG nova.compute.manager [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] No waiting events found dispatching network-vif-plugged-67d6315d-2ed0-45b1-b08e-d4967d01eae3 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.849034] env[61995]: WARNING nova.compute.manager [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Received unexpected event network-vif-plugged-67d6315d-2ed0-45b1-b08e-d4967d01eae3 for instance with vm_state building and task_state spawning. [ 830.849034] env[61995]: DEBUG nova.compute.manager [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Received event network-changed-67d6315d-2ed0-45b1-b08e-d4967d01eae3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.849034] env[61995]: DEBUG nova.compute.manager [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Refreshing instance network info cache due to event network-changed-67d6315d-2ed0-45b1-b08e-d4967d01eae3. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.849034] env[61995]: DEBUG oslo_concurrency.lockutils [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] Acquiring lock "refresh_cache-451e226c-5fb7-4042-a706-3e58b0606546" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.889894] env[61995]: DEBUG nova.network.neutron [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.001537] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "7524c912-4d73-462a-949f-3b4c85874094" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.365s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.091689] env[61995]: DEBUG nova.network.neutron [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Updating instance_info_cache with network_info: [{"id": "67d6315d-2ed0-45b1-b08e-d4967d01eae3", "address": "fa:16:3e:29:be:d2", "network": {"id": "dc52b688-31ca-43ac-a57a-34d823ee8cae", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-819260078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fb7e9f4ff4a4af8a216b3e38e864561", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a2b284a-a29c-478f-b763-c9b5821e20ec", "external-id": "nsx-vlan-transportzone-691", "segmentation_id": 691, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67d6315d-2e", "ovs_interfaceid": "67d6315d-2ed0-45b1-b08e-d4967d01eae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.103511] env[61995]: INFO nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Took 38.18 seconds to build instance. [ 831.252767] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4bda5ea5-3c72-4745-bb11-4df1e1fcd456 tempest-InstanceActionsNegativeTestJSON-1175069309 tempest-InstanceActionsNegativeTestJSON-1175069309-project-member] Lock "d056dadd-492d-4a4d-abc2-fc5f113f865b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.253s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.299191] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f7780b-635a-641d-599a-927e33af1354, 'name': SearchDatastore_Task, 'duration_secs': 0.021216} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.299482] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.299751] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] af7163e8-4e77-4336-80e9-0623e29ed8c0/af7163e8-4e77-4336-80e9-0623e29ed8c0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 831.300025] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a269dc98-0cde-49e4-ba60-4f3141340ee3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.307048] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 831.307048] env[61995]: value = "task-794580" [ 831.307048] env[61995]: _type = "Task" [ 831.307048] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.317937] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.332067] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794579, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.595706] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Releasing lock "refresh_cache-451e226c-5fb7-4042-a706-3e58b0606546" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.595706] env[61995]: DEBUG nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Instance network_info: |[{"id": "67d6315d-2ed0-45b1-b08e-d4967d01eae3", "address": "fa:16:3e:29:be:d2", "network": {"id": "dc52b688-31ca-43ac-a57a-34d823ee8cae", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-819260078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fb7e9f4ff4a4af8a216b3e38e864561", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a2b284a-a29c-478f-b763-c9b5821e20ec", "external-id": "nsx-vlan-transportzone-691", "segmentation_id": 691, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67d6315d-2e", "ovs_interfaceid": "67d6315d-2ed0-45b1-b08e-d4967d01eae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.596013] env[61995]: DEBUG oslo_concurrency.lockutils [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] Acquired lock "refresh_cache-451e226c-5fb7-4042-a706-3e58b0606546" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.596056] env[61995]: DEBUG nova.network.neutron [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Refreshing network info cache for port 67d6315d-2ed0-45b1-b08e-d4967d01eae3 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 831.597378] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:be:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a2b284a-a29c-478f-b763-c9b5821e20ec', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '67d6315d-2ed0-45b1-b08e-d4967d01eae3', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.607086] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Creating folder: Project (4fb7e9f4ff4a4af8a216b3e38e864561). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 831.608834] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.943s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.609101] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ecaa826-3eeb-4ddd-bac1-5d4b8973d3c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.622234] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Created folder: Project (4fb7e9f4ff4a4af8a216b3e38e864561) in parent group-v185203. [ 831.622459] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Creating folder: Instances. Parent ref: group-v185312. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 831.622713] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f5fa1dd-b9cb-47d5-8013-b157a1884ce5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.632345] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Created folder: Instances in parent group-v185312. [ 831.632597] env[61995]: DEBUG oslo.service.loopingcall [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.632798] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 831.633032] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be44b666-98ac-4366-b2e3-7c648eb69e60 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.654213] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.654213] env[61995]: value = "task-794583" [ 831.654213] env[61995]: _type = "Task" [ 831.654213] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.663575] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794583, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.819564] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482475} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.819844] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] af7163e8-4e77-4336-80e9-0623e29ed8c0/af7163e8-4e77-4336-80e9-0623e29ed8c0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 831.820085] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.820347] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab886a1f-b3da-4d49-af64-4bb508e4b0bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.830735] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 831.830735] env[61995]: value = "task-794584" [ 831.830735] env[61995]: _type = "Task" [ 831.830735] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.837458] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794579, 'name': CreateVM_Task, 'duration_secs': 0.532654} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.838064] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.838552] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.838721] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.839101] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.841830] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdb2bc89-3d16-42e9-90c5-b41197b9277f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.846791] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794584, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.849739] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 831.849739] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52243229-8ee1-15eb-b452-929058ed34c2" [ 831.849739] env[61995]: _type = "Task" [ 831.849739] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.859423] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52243229-8ee1-15eb-b452-929058ed34c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.098392] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf75f85-87e2-4744-86aa-d8945f6c8537 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.106363] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ac6516-0d86-41b4-9134-492f620fb6dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.143173] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc35d22-10bf-41f9-a1ba-b7d5b52f0127 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.151796] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780c41bd-1b69-41fe-9bb8-49cf3439be34 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.163995] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794583, 'name': CreateVM_Task, 'duration_secs': 0.416501} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.171779] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 832.172324] env[61995]: DEBUG nova.compute.provider_tree [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.177340] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.344150] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794584, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.160318} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.344150] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.344384] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-848a84dd-b80b-405a-b779-656ffd7ab709 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.367525] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] af7163e8-4e77-4336-80e9-0623e29ed8c0/af7163e8-4e77-4336-80e9-0623e29ed8c0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.370872] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81226af4-ab28-440e-a0ea-0ed2d619b857 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.391094] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52243229-8ee1-15eb-b452-929058ed34c2, 'name': SearchDatastore_Task, 'duration_secs': 0.008698} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.394314] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.394314] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.394314] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.394314] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.394314] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.394314] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 832.394314] env[61995]: value = "task-794585" [ 832.394314] env[61995]: _type = "Task" [ 832.394314] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.394314] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.394314] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.394948] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8afe87d-6882-442c-81e0-36edea483952 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.397024] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5524bc84-b444-4a4d-b4dc-10713fc700c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.405593] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 832.405593] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5287feb6-080d-7138-bed3-2021c9cbb4dc" [ 832.405593] env[61995]: _type = "Task" [ 832.405593] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.409904] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794585, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.415612] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.415870] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.416604] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d96e0b01-04d2-48b5-97f0-1deeb692efdd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.426527] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5287feb6-080d-7138-bed3-2021c9cbb4dc, 'name': SearchDatastore_Task, 'duration_secs': 0.011744} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.428104] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.428407] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.428693] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.429056] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 832.429056] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e098fb-ba94-9bef-b8cf-6174a0f4f44a" [ 832.429056] env[61995]: _type = "Task" [ 832.429056] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.438383] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e098fb-ba94-9bef-b8cf-6174a0f4f44a, 'name': SearchDatastore_Task, 'duration_secs': 0.011957} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.439038] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7356c446-1f5e-44d5-a9ad-b171993218d1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.444440] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 832.444440] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a9c591-f81c-a613-c9f2-34fbef11d588" [ 832.444440] env[61995]: _type = "Task" [ 832.444440] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.452776] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a9c591-f81c-a613-c9f2-34fbef11d588, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.636472] env[61995]: DEBUG nova.network.neutron [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Updated VIF entry in instance network info cache for port 67d6315d-2ed0-45b1-b08e-d4967d01eae3. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 832.636859] env[61995]: DEBUG nova.network.neutron [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Updating instance_info_cache with network_info: [{"id": "67d6315d-2ed0-45b1-b08e-d4967d01eae3", "address": "fa:16:3e:29:be:d2", "network": {"id": "dc52b688-31ca-43ac-a57a-34d823ee8cae", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-819260078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fb7e9f4ff4a4af8a216b3e38e864561", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a2b284a-a29c-478f-b763-c9b5821e20ec", "external-id": "nsx-vlan-transportzone-691", "segmentation_id": 691, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67d6315d-2e", "ovs_interfaceid": "67d6315d-2ed0-45b1-b08e-d4967d01eae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.676236] env[61995]: DEBUG nova.scheduler.client.report [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.908698] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794585, 'name': ReconfigVM_Task, 'duration_secs': 0.291732} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.909015] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Reconfigured VM instance instance-0000003a to attach disk [datastore1] af7163e8-4e77-4336-80e9-0623e29ed8c0/af7163e8-4e77-4336-80e9-0623e29ed8c0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.909624] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6655e31-b4a9-454d-9af4-e82a37360efc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.917595] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 832.917595] env[61995]: value = "task-794586" [ 832.917595] env[61995]: _type = "Task" [ 832.917595] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.925311] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794586, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.955435] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a9c591-f81c-a613-c9f2-34fbef11d588, 'name': SearchDatastore_Task, 'duration_secs': 0.012582} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.956380] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.956380] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 832.956380] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.956550] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.956708] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82ed54f5-d7dd-43a0-a314-ce7ab19e833f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.958847] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-098bdc77-52eb-4504-a310-3e10766061a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.969130] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 832.969130] env[61995]: value = "task-794587" [ 832.969130] env[61995]: _type = "Task" [ 832.969130] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.970502] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.970716] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.977134] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12872c85-dbd2-4b05-b441-acb3f7d622f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.985023] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.986428] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 832.986428] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525d4a6a-366e-1ac3-1a64-a5e515a11f29" [ 832.986428] env[61995]: _type = "Task" [ 832.986428] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.994884] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525d4a6a-366e-1ac3-1a64-a5e515a11f29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.143133] env[61995]: DEBUG oslo_concurrency.lockutils [req-360c6da2-b930-4cf6-ade1-545165a643b8 req-b9bf2150-b171-49ea-a0f8-dcebcfa70cff service nova] Releasing lock "refresh_cache-451e226c-5fb7-4042-a706-3e58b0606546" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.181238] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.181860] env[61995]: DEBUG nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.185367] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.420s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.185603] env[61995]: DEBUG nova.objects.instance [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lazy-loading 'resources' on Instance uuid 55555deb-9883-483b-a901-42b1a66c60d6 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.428779] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794586, 'name': Rename_Task, 'duration_secs': 0.148192} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.429134] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 833.429493] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40baa01b-9695-473a-b521-34b2a1589377 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.439729] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 833.439729] env[61995]: value = "task-794588" [ 833.439729] env[61995]: _type = "Task" [ 833.439729] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.449736] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794588, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.480726] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794587, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.498676] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525d4a6a-366e-1ac3-1a64-a5e515a11f29, 'name': SearchDatastore_Task, 'duration_secs': 0.019425} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.499720] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35df8cc9-4dce-48b0-8d27-1e9ce7f55b80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.506213] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 833.506213] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523c62a2-185f-8a2a-1752-654579bdf4aa" [ 833.506213] env[61995]: _type = "Task" [ 833.506213] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.516518] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523c62a2-185f-8a2a-1752-654579bdf4aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.689984] env[61995]: DEBUG nova.compute.utils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.694312] env[61995]: DEBUG nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.694460] env[61995]: DEBUG nova.network.neutron [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 833.769132] env[61995]: DEBUG nova.policy [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16bf12381242949167ffc13d3aa404', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e0840f3a4d24ea78c35c4b273f1e413', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 833.950915] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794588, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.983257] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637855} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.983257] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 833.984274] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.984274] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9042f766-8ba9-4f19-835e-604a288ea396 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.991181] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 833.991181] env[61995]: value = "task-794589" [ 833.991181] env[61995]: _type = "Task" [ 833.991181] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.005426] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794589, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.018804] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523c62a2-185f-8a2a-1752-654579bdf4aa, 'name': SearchDatastore_Task, 'duration_secs': 0.042697} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.020093] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.020093] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 451e226c-5fb7-4042-a706-3e58b0606546/451e226c-5fb7-4042-a706-3e58b0606546.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 834.020093] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0b33db4-8f81-4fd7-a392-3668c2066d11 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.026990] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 834.026990] env[61995]: value = "task-794590" [ 834.026990] env[61995]: _type = "Task" [ 834.026990] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.039022] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794590, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.083192] env[61995]: DEBUG nova.network.neutron [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Successfully created port: 9f82eee3-593f-46f6-a461-4c7aec380a4e {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.112019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a7f21c-50a3-47d2-8dbe-7de0a58c2c66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.119700] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da4b09d-3df9-4df1-bce4-e638ccf39bd2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.149138] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6793b8-1e59-4a0f-8914-0c5b93f881fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.157303] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7b5a1d-369d-49f3-b5f9-452adde36dc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.171134] env[61995]: DEBUG nova.compute.provider_tree [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.197826] env[61995]: DEBUG nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.454340] env[61995]: DEBUG oslo_vmware.api [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794588, 'name': PowerOnVM_Task, 'duration_secs': 0.593253} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.454602] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 834.454802] env[61995]: INFO nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Took 8.83 seconds to spawn the instance on the hypervisor. [ 834.455017] env[61995]: DEBUG nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.455783] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862cfbeb-a364-4cf5-bd6e-6fe60031912d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.503341] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794589, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081227} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.504364] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.505218] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36642e1-5eec-4ab3-9565-10b1930d4be0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.526487] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.527136] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b686e89-324e-4b68-96db-b8eeac7d5f9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.553841] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794590, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.555365] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 834.555365] env[61995]: value = "task-794591" [ 834.555365] env[61995]: _type = "Task" [ 834.555365] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.566321] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794591, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.677313] env[61995]: DEBUG nova.scheduler.client.report [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.972635] env[61995]: INFO nova.compute.manager [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Took 36.48 seconds to build instance. [ 835.056437] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.056999] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.064787] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794590, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.888568} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.076428] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 451e226c-5fb7-4042-a706-3e58b0606546/451e226c-5fb7-4042-a706-3e58b0606546.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 835.076428] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.076428] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d54ab1f-80fc-473f-956c-62901225b988 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.082138] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794591, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.084366] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 835.084366] env[61995]: value = "task-794592" [ 835.084366] env[61995]: _type = "Task" [ 835.084366] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.094473] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794592, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.182729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.997s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.185467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.689s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.185708] env[61995]: DEBUG nova.objects.instance [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lazy-loading 'resources' on Instance uuid 0d34066f-5b8f-4bac-9b83-67d78987c4b3 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.206029] env[61995]: INFO nova.scheduler.client.report [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Deleted allocations for instance 55555deb-9883-483b-a901-42b1a66c60d6 [ 835.208806] env[61995]: DEBUG nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.237363] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.237627] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.237790] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.238031] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.238209] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.238377] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.238596] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.238755] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.238922] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.239183] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.239403] env[61995]: DEBUG nova.virt.hardware [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.240261] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d1e57f-0540-4043-9d51-425a24a1d7d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.248376] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160e2cd0-54dd-45d4-a799-bedd352d6802 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.474568] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4620130f-29b8-4626-a260-cf6a955555b2 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.771s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.569800] env[61995]: DEBUG nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.579045] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794591, 'name': ReconfigVM_Task, 'duration_secs': 0.854299} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.579491] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.581086] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-369f0512-7647-4ef5-ae84-845573a12781 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.591907] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 835.591907] env[61995]: value = "task-794593" [ 835.591907] env[61995]: _type = "Task" [ 835.591907] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.603325] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794592, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066131} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.603786] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.604524] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9eedf9-fd8e-412b-b17a-22c8ee0ae7cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.610911] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794593, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.631656] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 451e226c-5fb7-4042-a706-3e58b0606546/451e226c-5fb7-4042-a706-3e58b0606546.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.632750] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2170f23-1d73-425d-a83d-19720299f6e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.654831] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 835.654831] env[61995]: value = "task-794594" [ 835.654831] env[61995]: _type = "Task" [ 835.654831] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.667448] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.722275] env[61995]: DEBUG oslo_concurrency.lockutils [None req-60153a0e-7248-42dd-82ef-6fc9cd7f85ae tempest-ServerShowV247Test-651911540 tempest-ServerShowV247Test-651911540-project-member] Lock "55555deb-9883-483b-a901-42b1a66c60d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.744s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.909696] env[61995]: DEBUG nova.compute.manager [req-cf211a0d-a50e-4c07-a859-be2f1a488b98 req-d6b3fc00-35d8-4244-9127-ed6a3c7bed5c service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Received event network-vif-plugged-9f82eee3-593f-46f6-a461-4c7aec380a4e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.911359] env[61995]: DEBUG oslo_concurrency.lockutils [req-cf211a0d-a50e-4c07-a859-be2f1a488b98 req-d6b3fc00-35d8-4244-9127-ed6a3c7bed5c service nova] Acquiring lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.911910] env[61995]: DEBUG oslo_concurrency.lockutils [req-cf211a0d-a50e-4c07-a859-be2f1a488b98 req-d6b3fc00-35d8-4244-9127-ed6a3c7bed5c service nova] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.912341] env[61995]: DEBUG oslo_concurrency.lockutils [req-cf211a0d-a50e-4c07-a859-be2f1a488b98 req-d6b3fc00-35d8-4244-9127-ed6a3c7bed5c service nova] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.912770] env[61995]: DEBUG nova.compute.manager [req-cf211a0d-a50e-4c07-a859-be2f1a488b98 req-d6b3fc00-35d8-4244-9127-ed6a3c7bed5c service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] No waiting events found dispatching network-vif-plugged-9f82eee3-593f-46f6-a461-4c7aec380a4e {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.913118] env[61995]: WARNING nova.compute.manager [req-cf211a0d-a50e-4c07-a859-be2f1a488b98 req-d6b3fc00-35d8-4244-9127-ed6a3c7bed5c service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Received unexpected event network-vif-plugged-9f82eee3-593f-46f6-a461-4c7aec380a4e for instance with vm_state building and task_state spawning. [ 835.984128] env[61995]: DEBUG nova.network.neutron [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Successfully updated port: 9f82eee3-593f-46f6-a461-4c7aec380a4e {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.098716] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.106527] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794593, 'name': Rename_Task, 'duration_secs': 0.316333} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.106527] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.106527] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-427fe5b5-0d9f-4c1a-8f84-758295ed32d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.112152] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 836.112152] env[61995]: value = "task-794595" [ 836.112152] env[61995]: _type = "Task" [ 836.112152] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.120323] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.124996] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167d206b-a31a-4378-8403-16b5ba399017 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.132159] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48383982-5e10-47e4-b12c-adf45f17b92b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.171921] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b95250-2737-4291-b153-f37ae5c56ca8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.179918] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794594, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.184044] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6bb2e1f-ae0f-4100-91dc-073386378cad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.197720] env[61995]: DEBUG nova.compute.provider_tree [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.323693] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "7524c912-4d73-462a-949f-3b4c85874094" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.324186] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "7524c912-4d73-462a-949f-3b4c85874094" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.324255] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "7524c912-4d73-462a-949f-3b4c85874094-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.324435] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "7524c912-4d73-462a-949f-3b4c85874094-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.324783] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "7524c912-4d73-462a-949f-3b4c85874094-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.327652] env[61995]: INFO nova.compute.manager [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Terminating instance [ 836.330025] env[61995]: DEBUG nova.compute.manager [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.330110] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 836.331113] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ea4886-4c5d-4753-97b2-99c41a6c7869 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.339528] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.339847] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b36ca3f1-db3f-48f1-a88e-e45507ccc1cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.347027] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 836.347027] env[61995]: value = "task-794596" [ 836.347027] env[61995]: _type = "Task" [ 836.347027] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.356537] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.485747] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "refresh_cache-d3f629c5-96f4-4208-be0d-cbf9810ecc6b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.486137] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "refresh_cache-d3f629c5-96f4-4208-be0d-cbf9810ecc6b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.486175] env[61995]: DEBUG nova.network.neutron [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.624502] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794595, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.678529] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794594, 'name': ReconfigVM_Task, 'duration_secs': 0.717166} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.678529] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 451e226c-5fb7-4042-a706-3e58b0606546/451e226c-5fb7-4042-a706-3e58b0606546.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.680492] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be4b7bf7-3074-447f-94b7-b908ed2bb275 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.685596] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 836.685596] env[61995]: value = "task-794597" [ 836.685596] env[61995]: _type = "Task" [ 836.685596] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.693646] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794597, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.700789] env[61995]: DEBUG nova.scheduler.client.report [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.857308] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.019986] env[61995]: DEBUG nova.network.neutron [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 837.130634] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794595, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.202844] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794597, 'name': Rename_Task, 'duration_secs': 0.17491} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.203253] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 837.203565] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44c27393-66c6-4ac8-8c34-51e9d6665ff0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.206096] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.021s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.208422] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 20.428s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.217809] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 837.217809] env[61995]: value = "task-794598" [ 837.217809] env[61995]: _type = "Task" [ 837.217809] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.237745] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794598, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.239082] env[61995]: INFO nova.scheduler.client.report [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Deleted allocations for instance 0d34066f-5b8f-4bac-9b83-67d78987c4b3 [ 837.265105] env[61995]: DEBUG nova.network.neutron [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Updating instance_info_cache with network_info: [{"id": "9f82eee3-593f-46f6-a461-4c7aec380a4e", "address": "fa:16:3e:40:42:e1", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f82eee3-59", "ovs_interfaceid": "9f82eee3-593f-46f6-a461-4c7aec380a4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.359352] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794596, 'name': PowerOffVM_Task, 'duration_secs': 0.775676} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.359694] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 837.359862] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 837.360161] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3318ac2-3f4d-4f1c-a718-cf22d59d36c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.467563] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 837.467829] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 837.468100] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleting the datastore file [datastore1] 7524c912-4d73-462a-949f-3b4c85874094 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.468397] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c095250f-87b8-4fc8-bcfd-a940270d822b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.475308] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 837.475308] env[61995]: value = "task-794600" [ 837.475308] env[61995]: _type = "Task" [ 837.475308] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.485263] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.624736] env[61995]: DEBUG oslo_vmware.api [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794595, 'name': PowerOnVM_Task, 'duration_secs': 1.279429} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.625129] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.625308] env[61995]: INFO nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Took 6.92 seconds to spawn the instance on the hypervisor. [ 837.625589] env[61995]: DEBUG nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.626310] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2394192b-048e-4150-8fd7-45bf2f1bfd93 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.737947] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794598, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.748150] env[61995]: DEBUG oslo_concurrency.lockutils [None req-485a8624-cee4-458e-b1b9-ec0bfe37845a tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "0d34066f-5b8f-4bac-9b83-67d78987c4b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.390s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.770292] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "refresh_cache-d3f629c5-96f4-4208-be0d-cbf9810ecc6b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.770292] env[61995]: DEBUG nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Instance network_info: |[{"id": "9f82eee3-593f-46f6-a461-4c7aec380a4e", "address": "fa:16:3e:40:42:e1", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f82eee3-59", "ovs_interfaceid": "9f82eee3-593f-46f6-a461-4c7aec380a4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.770292] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:42:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff1f3320-df8e-49df-a412-9797a23bd173', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f82eee3-593f-46f6-a461-4c7aec380a4e', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.777758] env[61995]: DEBUG oslo.service.loopingcall [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.783094] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.783094] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7eded748-9170-4534-b968-51af7b4e2c4e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.808342] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.808342] env[61995]: value = "task-794601" [ 837.808342] env[61995]: _type = "Task" [ 837.808342] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.816289] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794601, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.991791] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.006469] env[61995]: DEBUG nova.compute.manager [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Received event network-changed-9f82eee3-593f-46f6-a461-4c7aec380a4e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 838.006674] env[61995]: DEBUG nova.compute.manager [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Refreshing instance network info cache due to event network-changed-9f82eee3-593f-46f6-a461-4c7aec380a4e. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 838.006893] env[61995]: DEBUG oslo_concurrency.lockutils [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] Acquiring lock "refresh_cache-d3f629c5-96f4-4208-be0d-cbf9810ecc6b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.007056] env[61995]: DEBUG oslo_concurrency.lockutils [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] Acquired lock "refresh_cache-d3f629c5-96f4-4208-be0d-cbf9810ecc6b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.007228] env[61995]: DEBUG nova.network.neutron [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Refreshing network info cache for port 9f82eee3-593f-46f6-a461-4c7aec380a4e {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.150850] env[61995]: INFO nova.compute.manager [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Took 36.57 seconds to build instance. [ 838.218864] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de599a6c-d1d9-445d-a7bb-235a70100250 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.237938] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794598, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.238935] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee304fb2-f097-4cdd-8a6a-722e5b5ca552 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.275266] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1b4537-1b55-4ecc-907c-5a34abb909cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.286371] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedb1c60-e207-400c-b832-d3a85461f22f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.301202] env[61995]: DEBUG nova.compute.provider_tree [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.319387] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794601, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.489413] env[61995]: DEBUG oslo_vmware.api [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794600, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.543409} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.489668] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.489859] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 838.490172] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 838.490314] env[61995]: INFO nova.compute.manager [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Took 2.16 seconds to destroy the instance on the hypervisor. [ 838.490564] env[61995]: DEBUG oslo.service.loopingcall [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.490773] env[61995]: DEBUG nova.compute.manager [-] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 838.490854] env[61995]: DEBUG nova.network.neutron [-] [instance: 7524c912-4d73-462a-949f-3b4c85874094] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 838.514159] env[61995]: INFO nova.compute.manager [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Rebuilding instance [ 838.581129] env[61995]: DEBUG nova.compute.manager [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.581129] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40878be0-0522-4c20-b4c6-09c971c7498d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.653451] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6722e953-5392-48c0-a549-42b5baefce82 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.326s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.748473] env[61995]: DEBUG oslo_vmware.api [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794598, 'name': PowerOnVM_Task, 'duration_secs': 1.103791} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.748750] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 838.750031] env[61995]: INFO nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Took 9.57 seconds to spawn the instance on the hypervisor. [ 838.750031] env[61995]: DEBUG nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.750031] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b257ec5d-6239-4dd4-bb3e-ab9bc3fec03d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.805200] env[61995]: DEBUG nova.scheduler.client.report [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.829147] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794601, 'name': CreateVM_Task, 'duration_secs': 0.516944} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.829446] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.830082] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.830273] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.830653] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.831188] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a66ae105-4bf3-4b86-bc72-2e4d70712295 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.838391] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 838.838391] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526f63a6-6f2e-e705-a269-feab664043c7" [ 838.838391] env[61995]: _type = "Task" [ 838.838391] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.848700] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526f63a6-6f2e-e705-a269-feab664043c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.033062] env[61995]: DEBUG nova.network.neutron [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Updated VIF entry in instance network info cache for port 9f82eee3-593f-46f6-a461-4c7aec380a4e. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 839.033195] env[61995]: DEBUG nova.network.neutron [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Updating instance_info_cache with network_info: [{"id": "9f82eee3-593f-46f6-a461-4c7aec380a4e", "address": "fa:16:3e:40:42:e1", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f82eee3-59", "ovs_interfaceid": "9f82eee3-593f-46f6-a461-4c7aec380a4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.093519] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.093910] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6e9b87b-01d2-4c1a-bb18-10f3e34a315e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.101832] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 839.101832] env[61995]: value = "task-794602" [ 839.101832] env[61995]: _type = "Task" [ 839.101832] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.118670] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794602, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.242745] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "e53d10a3-1d16-498d-b4d8-abe7205410ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.243044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.243301] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "e53d10a3-1d16-498d-b4d8-abe7205410ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.243460] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.243637] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.247532] env[61995]: INFO nova.compute.manager [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Terminating instance [ 839.250035] env[61995]: DEBUG nova.compute.manager [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.250173] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.251060] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a812e4ed-7f86-4f67-9a9c-de978770e705 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.259599] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.260148] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12dbb7de-406c-4000-8d5a-23fe7ec86010 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.270372] env[61995]: DEBUG oslo_vmware.api [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 839.270372] env[61995]: value = "task-794603" [ 839.270372] env[61995]: _type = "Task" [ 839.270372] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.278234] env[61995]: INFO nova.compute.manager [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Took 38.23 seconds to build instance. [ 839.284308] env[61995]: DEBUG oslo_vmware.api [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.351019] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526f63a6-6f2e-e705-a269-feab664043c7, 'name': SearchDatastore_Task, 'duration_secs': 0.037907} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.351019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.351019] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.351019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.351019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.351019] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.351275] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3be47173-d9a7-49d3-abaa-632c96fa7b16 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.362378] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.362378] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 839.362378] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5c0dfd2-ad8a-4a9a-a8a8-1e8a948063bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.367168] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 839.367168] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c55eb0-269d-091a-d26c-31daeab3a133" [ 839.367168] env[61995]: _type = "Task" [ 839.367168] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.376084] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c55eb0-269d-091a-d26c-31daeab3a133, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.504474] env[61995]: DEBUG nova.network.neutron [-] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.538484] env[61995]: DEBUG oslo_concurrency.lockutils [req-301d5e32-8afa-4eea-8027-bcf56273fd31 req-395da0b1-1a5a-47dc-b8d6-bfb6c4be0d00 service nova] Releasing lock "refresh_cache-d3f629c5-96f4-4208-be0d-cbf9810ecc6b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.614653] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794602, 'name': PowerOffVM_Task, 'duration_secs': 0.160489} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.614653] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.615921] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.615921] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c087d5de-fafe-4c1a-b677-ffc9e89f2936 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.623382] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.623965] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0a4ead6-02ec-42ec-b8eb-b9350ca8dba4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.651121] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.651377] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.651575] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Deleting the datastore file [datastore2] 966a72d8-a51a-44f9-8184-8108f8cc3ce6 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.651866] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15eed90a-c200-4c4f-b34e-9963c422b4b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.663730] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 839.663730] env[61995]: value = "task-794605" [ 839.663730] env[61995]: _type = "Task" [ 839.663730] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.677123] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.687847] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "451e226c-5fb7-4042-a706-3e58b0606546" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.779254] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a98e7d43-5b9d-4a4c-85e0-affebf6441b9 tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "451e226c-5fb7-4042-a706-3e58b0606546" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.605s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.781427] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "451e226c-5fb7-4042-a706-3e58b0606546" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.094s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.781662] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "451e226c-5fb7-4042-a706-3e58b0606546-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.782100] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "451e226c-5fb7-4042-a706-3e58b0606546-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.782175] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "451e226c-5fb7-4042-a706-3e58b0606546-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.786880] env[61995]: DEBUG oslo_vmware.api [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794603, 'name': PowerOffVM_Task, 'duration_secs': 0.215293} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.787643] env[61995]: INFO nova.compute.manager [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Terminating instance [ 839.788990] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.789573] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.789849] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f6cff7d-5be3-45ad-a478-bacb15fe7ed6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.792877] env[61995]: DEBUG nova.compute.manager [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.792877] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.796314] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497c760f-62bc-4af0-af65-546416816e66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.806142] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.806142] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9a2843b-3ca8-497b-a826-4d07a22c404a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.810867] env[61995]: DEBUG oslo_vmware.api [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 839.810867] env[61995]: value = "task-794607" [ 839.810867] env[61995]: _type = "Task" [ 839.810867] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.816094] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.608s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.819517] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.646s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.819728] env[61995]: DEBUG nova.objects.instance [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 839.834108] env[61995]: DEBUG oslo_vmware.api [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.864888] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.865150] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.865340] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Deleting the datastore file [datastore2] e53d10a3-1d16-498d-b4d8-abe7205410ec {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.865610] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3506d4a-9152-4e6c-8869-d51d2de7c163 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.883394] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c55eb0-269d-091a-d26c-31daeab3a133, 'name': SearchDatastore_Task, 'duration_secs': 0.014288} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.886804] env[61995]: DEBUG oslo_vmware.api [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for the task: (returnval){ [ 839.886804] env[61995]: value = "task-794608" [ 839.886804] env[61995]: _type = "Task" [ 839.886804] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.886804] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e195a8c-b874-4753-bd79-052688a67c38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.897584] env[61995]: DEBUG oslo_vmware.api [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794608, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.901196] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 839.901196] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a38f63-c623-bed8-5913-2a36c7fe1219" [ 839.901196] env[61995]: _type = "Task" [ 839.901196] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.914559] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a38f63-c623-bed8-5913-2a36c7fe1219, 'name': SearchDatastore_Task, 'duration_secs': 0.012221} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.914892] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.915199] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] d3f629c5-96f4-4208-be0d-cbf9810ecc6b/d3f629c5-96f4-4208-be0d-cbf9810ecc6b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.915485] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c3ffd63-490c-47aa-b83e-483769e6a987 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.922936] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 839.922936] env[61995]: value = "task-794609" [ 839.922936] env[61995]: _type = "Task" [ 839.922936] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.932047] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.007486] env[61995]: INFO nova.compute.manager [-] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Took 1.52 seconds to deallocate network for instance. [ 840.039159] env[61995]: DEBUG nova.compute.manager [req-4f66dad3-b780-4a3d-983a-af5d480be94b req-5feb0845-8841-4e4b-bf73-503afaa47dd0 service nova] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Received event network-vif-deleted-2f7661f6-7c37-41c5-97d4-282602f6df05 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 840.175173] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228896} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.178150] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.178150] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.178150] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.323348] env[61995]: DEBUG oslo_vmware.api [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794607, 'name': PowerOffVM_Task, 'duration_secs': 0.233992} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.323348] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.323522] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.323702] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9eafc3d0-787a-41b2-ad8e-50f412e9d5e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.387089] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "ed952baf-f1ef-4ec0-8a15-928287190456" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.387870] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "ed952baf-f1ef-4ec0-8a15-928287190456" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.399036] env[61995]: DEBUG oslo_vmware.api [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Task: {'id': task-794608, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17272} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.399333] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.399530] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.399715] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.399895] env[61995]: INFO nova.compute.manager [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Took 1.15 seconds to destroy the instance on the hypervisor. [ 840.400589] env[61995]: DEBUG oslo.service.loopingcall [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.401715] env[61995]: DEBUG nova.compute.manager [-] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.401715] env[61995]: DEBUG nova.network.neutron [-] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.404527] env[61995]: INFO nova.scheduler.client.report [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Deleted allocation for migration e608fd92-962e-49ec-8fe1-0aeb1cf2b857 [ 840.411644] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.411644] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.411644] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Deleting the datastore file [datastore2] 451e226c-5fb7-4042-a706-3e58b0606546 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.411923] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14228901-7c16-44b2-b8f1-814fa6bcc46a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.419025] env[61995]: DEBUG oslo_vmware.api [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for the task: (returnval){ [ 840.419025] env[61995]: value = "task-794611" [ 840.419025] env[61995]: _type = "Task" [ 840.419025] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.428821] env[61995]: DEBUG oslo_vmware.api [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.435660] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491919} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.435915] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] d3f629c5-96f4-4208-be0d-cbf9810ecc6b/d3f629c5-96f4-4208-be0d-cbf9810ecc6b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.436162] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.436438] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6250b11d-bbc7-4eb4-bf8b-e3035c74ebd5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.447019] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 840.447019] env[61995]: value = "task-794612" [ 840.447019] env[61995]: _type = "Task" [ 840.447019] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.457216] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794612, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.517198] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.782839] env[61995]: DEBUG nova.objects.instance [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lazy-loading 'flavor' on Instance uuid f9c1424f-6096-481c-aa91-acc25bfd8f6e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.842023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0f4bd7fe-b29e-4939-ac83-a686a7985d45 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.842023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.611s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.842023] env[61995]: DEBUG nova.objects.instance [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lazy-loading 'resources' on Instance uuid 4da426a1-7af9-4e4e-82ed-c45687a19b86 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.890876] env[61995]: DEBUG nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.913014] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6786c560-3071-46c6-9b93-ad9e5434c8f7 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 27.703s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.933310] env[61995]: DEBUG oslo_vmware.api [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.958317] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794612, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.500286} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.958722] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.959851] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4444d73c-0187-4935-9eb8-56efc01df642 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.986979] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] d3f629c5-96f4-4208-be0d-cbf9810ecc6b/d3f629c5-96f4-4208-be0d-cbf9810ecc6b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.986979] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8443985b-8842-42f3-a81c-84f9bbc12d37 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.010385] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 841.010385] env[61995]: value = "task-794613" [ 841.010385] env[61995]: _type = "Task" [ 841.010385] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.021016] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.213730] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.214102] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.214230] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.215676] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.215676] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.215676] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.215676] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.215676] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.215856] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.215984] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.216240] env[61995]: DEBUG nova.virt.hardware [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.217466] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289d19f7-a174-4708-a341-dc0efc691a6e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.231464] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4deb5500-7840-4711-a154-a347a99a0286 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.245576] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.251335] env[61995]: DEBUG oslo.service.loopingcall [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.251335] env[61995]: DEBUG nova.network.neutron [-] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.253371] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.253723] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9c72f59-8109-445b-b9f3-276e5ff350ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.276746] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.276746] env[61995]: value = "task-794614" [ 841.276746] env[61995]: _type = "Task" [ 841.276746] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.287842] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794614, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.288372] env[61995]: DEBUG oslo_concurrency.lockutils [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.288573] env[61995]: DEBUG oslo_concurrency.lockutils [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquired lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.431763] env[61995]: DEBUG oslo_vmware.api [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Task: {'id': task-794611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.614524} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.431763] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.431763] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 841.431763] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 841.431763] env[61995]: INFO nova.compute.manager [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Took 1.64 seconds to destroy the instance on the hypervisor. [ 841.431763] env[61995]: DEBUG oslo.service.loopingcall [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.431763] env[61995]: DEBUG nova.compute.manager [-] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.431763] env[61995]: DEBUG nova.network.neutron [-] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 841.456360] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.522100] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794613, 'name': ReconfigVM_Task, 'duration_secs': 0.498869} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.522604] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Reconfigured VM instance instance-0000003d to attach disk [datastore2] d3f629c5-96f4-4208-be0d-cbf9810ecc6b/d3f629c5-96f4-4208-be0d-cbf9810ecc6b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.523499] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62ea3e6b-48f3-427d-b63d-fc494cf0ef4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.535180] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 841.535180] env[61995]: value = "task-794615" [ 841.535180] env[61995]: _type = "Task" [ 841.535180] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.549375] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794615, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.774344] env[61995]: INFO nova.compute.manager [-] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Took 1.37 seconds to deallocate network for instance. [ 841.792672] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794614, 'name': CreateVM_Task, 'duration_secs': 0.393934} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.793114] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 841.793300] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.793465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.793861] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.794082] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54b0a360-4b8f-4e35-936e-82a05089f6e6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.809718] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 841.809718] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526e7d47-c44f-9c4f-9d5e-1687d87c058c" [ 841.809718] env[61995]: _type = "Task" [ 841.809718] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.819042] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526e7d47-c44f-9c4f-9d5e-1687d87c058c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.919792] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26944e81-648a-4017-aa47-f7b171580697 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.932802] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323e04c9-6a4d-4b23-bf2f-88bddad69e7f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.978116] env[61995]: DEBUG nova.network.neutron [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.981468] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f828f8-9d7b-4191-a112-e9cdbd5db2d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.992338] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09129f53-4243-4db0-a209-444fefd8beb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.006230] env[61995]: DEBUG nova.compute.provider_tree [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.049188] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794615, 'name': Rename_Task, 'duration_secs': 0.192372} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.049530] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 842.049867] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a087d968-bfe3-4738-a0cb-c34b5d3b5b3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.055912] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 842.055912] env[61995]: value = "task-794616" [ 842.055912] env[61995]: _type = "Task" [ 842.055912] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.064413] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.071941] env[61995]: DEBUG nova.compute.manager [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Received event network-vif-deleted-52c15b2e-a283-4a39-9a38-d2ccd7da93a2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.072187] env[61995]: DEBUG nova.compute.manager [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Received event network-vif-deleted-67d6315d-2ed0-45b1-b08e-d4967d01eae3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.072437] env[61995]: INFO nova.compute.manager [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Neutron deleted interface 67d6315d-2ed0-45b1-b08e-d4967d01eae3; detaching it from the instance and deleting it from the info cache [ 842.072565] env[61995]: DEBUG nova.network.neutron [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.287647] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.325334] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526e7d47-c44f-9c4f-9d5e-1687d87c058c, 'name': SearchDatastore_Task, 'duration_secs': 0.016726} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.325334] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.325334] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.325334] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.325334] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.326178] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.326178] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-112a7814-ae82-411c-a6f6-138446991767 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.337119] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.337119] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.337947] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-194ae395-cf5a-4e7e-bd07-df3abe5a2dd6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.343579] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 842.343579] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52884603-a491-eec8-dbf7-6130b4fad44b" [ 842.343579] env[61995]: _type = "Task" [ 842.343579] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.352404] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52884603-a491-eec8-dbf7-6130b4fad44b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.486065] env[61995]: DEBUG nova.network.neutron [-] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.509371] env[61995]: DEBUG nova.scheduler.client.report [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.566021] env[61995]: DEBUG oslo_vmware.api [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794616, 'name': PowerOnVM_Task, 'duration_secs': 0.479282} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.566021] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 842.566186] env[61995]: INFO nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Took 7.36 seconds to spawn the instance on the hypervisor. [ 842.566368] env[61995]: DEBUG nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.567193] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2042295-bda0-478a-ac74-5d44a000064b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.578386] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86d2657d-8b6e-4ff5-9224-5e3fb81c0987 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.586448] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7add155e-06fd-4a34-9a14-b2e430a078d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.618607] env[61995]: DEBUG nova.compute.manager [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Detach interface failed, port_id=67d6315d-2ed0-45b1-b08e-d4967d01eae3, reason: Instance 451e226c-5fb7-4042-a706-3e58b0606546 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 842.618849] env[61995]: DEBUG nova.compute.manager [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Received event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.619035] env[61995]: DEBUG nova.compute.manager [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing instance network info cache due to event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 842.619278] env[61995]: DEBUG oslo_concurrency.lockutils [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] Acquiring lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.853853] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52884603-a491-eec8-dbf7-6130b4fad44b, 'name': SearchDatastore_Task, 'duration_secs': 0.008878} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.854635] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9300c95d-b5ac-415a-a8b0-cd0ce4b1701c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.860362] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 842.860362] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529f7436-10ba-90eb-f018-8f62e8bb9471" [ 842.860362] env[61995]: _type = "Task" [ 842.860362] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.867933] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529f7436-10ba-90eb-f018-8f62e8bb9471, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.990984] env[61995]: INFO nova.compute.manager [-] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Took 1.56 seconds to deallocate network for instance. [ 843.016797] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.176s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.023030] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.479s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.023030] env[61995]: INFO nova.compute.claims [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.033870] env[61995]: DEBUG nova.network.neutron [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.064020] env[61995]: INFO nova.scheduler.client.report [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Deleted allocations for instance 4da426a1-7af9-4e4e-82ed-c45687a19b86 [ 843.091677] env[61995]: INFO nova.compute.manager [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Took 37.12 seconds to build instance. [ 843.180865] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.181273] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.373544] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529f7436-10ba-90eb-f018-8f62e8bb9471, 'name': SearchDatastore_Task, 'duration_secs': 0.027775} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.373544] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.373544] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.373544] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c616afb3-a613-46fa-a01c-d3e53a29d115 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.381035] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 843.381035] env[61995]: value = "task-794617" [ 843.381035] env[61995]: _type = "Task" [ 843.381035] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.391342] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.499138] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.539276] env[61995]: DEBUG oslo_concurrency.lockutils [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Releasing lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.539515] env[61995]: DEBUG nova.compute.manager [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Inject network info {{(pid=61995) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 843.539784] env[61995]: DEBUG nova.compute.manager [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] network_info to inject: |[{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 843.545774] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Reconfiguring VM instance to set the machine id {{(pid=61995) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 843.547015] env[61995]: DEBUG oslo_concurrency.lockutils [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] Acquired lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.547275] env[61995]: DEBUG nova.network.neutron [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 843.553015] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce4a7eb1-743e-433f-94ad-ff6988979df8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.573070] env[61995]: DEBUG oslo_vmware.api [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 843.573070] env[61995]: value = "task-794618" [ 843.573070] env[61995]: _type = "Task" [ 843.573070] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.574627] env[61995]: DEBUG oslo_concurrency.lockutils [None req-068c0d96-7ec8-43d2-93da-5e9a4ab7b3c1 tempest-ImagesOneServerTestJSON-1719202296 tempest-ImagesOneServerTestJSON-1719202296-project-member] Lock "4da426a1-7af9-4e4e-82ed-c45687a19b86" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.445s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.586228] env[61995]: DEBUG oslo_vmware.api [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794618, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.595076] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5513dd8f-499d-4188-938c-c2a95a89605a tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.891s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.684256] env[61995]: DEBUG nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.797916] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.798287] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.802269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.802269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.802269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.805503] env[61995]: INFO nova.compute.manager [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Terminating instance [ 843.807589] env[61995]: DEBUG nova.compute.manager [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.808044] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 843.809884] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caad4e45-8b77-4fa5-8122-8722ae10b9ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.821391] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 843.821391] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7632e24d-fa90-4fb7-9e63-68c520e0775b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.828992] env[61995]: DEBUG oslo_vmware.api [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 843.828992] env[61995]: value = "task-794619" [ 843.828992] env[61995]: _type = "Task" [ 843.828992] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.839810] env[61995]: DEBUG oslo_vmware.api [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.892167] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794617, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.093339] env[61995]: DEBUG oslo_vmware.api [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794618, 'name': ReconfigVM_Task, 'duration_secs': 0.203451} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.093419] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-358a8eb7-c995-4a28-8398-e96956c4ebb8 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Reconfigured VM instance to set the machine id {{(pid=61995) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 844.122354] env[61995]: DEBUG nova.objects.instance [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lazy-loading 'flavor' on Instance uuid f9c1424f-6096-481c-aa91-acc25bfd8f6e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.209517] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.302547] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "eef1cf32-e71e-4696-8170-1023e3a5518d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.302916] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.303156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "eef1cf32-e71e-4696-8170-1023e3a5518d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.303346] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.303516] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.305488] env[61995]: INFO nova.compute.manager [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Terminating instance [ 844.312409] env[61995]: DEBUG nova.compute.manager [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.312409] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 844.312827] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f9ca03-3f67-4d60-a274-f12190a44e13 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.320368] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 844.320610] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5faf12ce-9c0e-4bee-815c-ed4776504394 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.327178] env[61995]: DEBUG oslo_vmware.api [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 844.327178] env[61995]: value = "task-794620" [ 844.327178] env[61995]: _type = "Task" [ 844.327178] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.341076] env[61995]: DEBUG oslo_vmware.api [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.346370] env[61995]: DEBUG oslo_vmware.api [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794619, 'name': PowerOffVM_Task, 'duration_secs': 0.193198} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.353016] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 844.353016] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 844.353386] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c082ef51-7f67-4f6e-a7b4-50b09b8a6f4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.395683] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522322} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.395683] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 844.395881] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.398580] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63c09288-d899-4869-a43b-f24fb9dfaf55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.405638] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 844.405638] env[61995]: value = "task-794622" [ 844.405638] env[61995]: _type = "Task" [ 844.405638] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.413747] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.446806] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 844.447059] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 844.447297] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleting the datastore file [datastore2] d3f629c5-96f4-4208-be0d-cbf9810ecc6b {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.447772] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e3d4727-be13-4d3e-9e1c-7722f03b3f14 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.453778] env[61995]: DEBUG oslo_vmware.api [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 844.453778] env[61995]: value = "task-794623" [ 844.453778] env[61995]: _type = "Task" [ 844.453778] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.462272] env[61995]: DEBUG oslo_vmware.api [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794623, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.490104] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec963a13-8cf2-426b-b184-6cee9cee56c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.497445] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91df9e64-a510-42f9-8373-7c593340568a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.529236] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63cb476-e541-4ff5-b5fd-24a4ede2214b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.538412] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19790d9e-c832-4b79-a6ed-47cf0744410c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.543527] env[61995]: DEBUG nova.network.neutron [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updated VIF entry in instance network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 844.543989] env[61995]: DEBUG nova.network.neutron [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.559261] env[61995]: DEBUG nova.compute.provider_tree [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.627362] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.838504] env[61995]: DEBUG oslo_vmware.api [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794620, 'name': PowerOffVM_Task, 'duration_secs': 0.243497} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.838952] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 844.839289] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 844.839619] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d6ff375-a434-454f-8682-b5795adc90b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.921219] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066161} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.924089] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.927021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9779cf73-c9d5-4725-b31a-ce43e2a9e4bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.948118] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.949959] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9226f84-42f7-4f36-b505-e14c2e1faa31 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.964950] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 844.965368] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 844.965661] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Deleting the datastore file [datastore1] eef1cf32-e71e-4696-8170-1023e3a5518d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.966123] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73f669ae-8c1e-4745-9bb4-53ae2bda14a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.977610] env[61995]: DEBUG oslo_vmware.api [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794623, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159517} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.979891] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.980220] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 844.980510] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 844.980807] env[61995]: INFO nova.compute.manager [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 844.981156] env[61995]: DEBUG oslo.service.loopingcall [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.981508] env[61995]: DEBUG oslo_vmware.api [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for the task: (returnval){ [ 844.981508] env[61995]: value = "task-794625" [ 844.981508] env[61995]: _type = "Task" [ 844.981508] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.985016] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 844.985016] env[61995]: value = "task-794626" [ 844.985016] env[61995]: _type = "Task" [ 844.985016] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.985016] env[61995]: DEBUG nova.compute.manager [-] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.985016] env[61995]: DEBUG nova.network.neutron [-] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 844.994946] env[61995]: DEBUG oslo_vmware.api [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794625, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.998247] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794626, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.051030] env[61995]: DEBUG oslo_concurrency.lockutils [req-6675ef88-8228-4b80-a5d9-36df416074ef req-30eec4de-5fb1-4938-92fd-53b412a1d6e6 service nova] Releasing lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.051030] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquired lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.065228] env[61995]: DEBUG nova.scheduler.client.report [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.504059] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794626, 'name': ReconfigVM_Task, 'duration_secs': 0.279704} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.504059] env[61995]: DEBUG oslo_vmware.api [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Task: {'id': task-794625, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15305} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.504059] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 966a72d8-a51a-44f9-8184-8108f8cc3ce6/966a72d8-a51a-44f9-8184-8108f8cc3ce6.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.504059] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.504059] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 845.504059] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 845.504059] env[61995]: INFO nova.compute.manager [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Took 1.19 seconds to destroy the instance on the hypervisor. [ 845.504059] env[61995]: DEBUG oslo.service.loopingcall [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.504059] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8ed5718-e69d-4a19-9998-e6ade1639964 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.507144] env[61995]: DEBUG nova.compute.manager [-] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.507543] env[61995]: DEBUG nova.network.neutron [-] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 845.519119] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 845.519119] env[61995]: value = "task-794627" [ 845.519119] env[61995]: _type = "Task" [ 845.519119] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.528833] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794627, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.574018] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.574018] env[61995]: DEBUG nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.575337] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.983s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.575841] env[61995]: DEBUG nova.objects.instance [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lazy-loading 'resources' on Instance uuid 823d2169-2bcb-4d5a-9677-cc4ac49a7e84 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.678953] env[61995]: DEBUG nova.compute.manager [req-5fbdde08-31bd-4078-9959-1be0d259342a req-017426b5-1bf3-4c2b-a660-c998100f4002 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Received event network-vif-deleted-9f82eee3-593f-46f6-a461-4c7aec380a4e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 845.679696] env[61995]: INFO nova.compute.manager [req-5fbdde08-31bd-4078-9959-1be0d259342a req-017426b5-1bf3-4c2b-a660-c998100f4002 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Neutron deleted interface 9f82eee3-593f-46f6-a461-4c7aec380a4e; detaching it from the instance and deleting it from the info cache [ 845.679696] env[61995]: DEBUG nova.network.neutron [req-5fbdde08-31bd-4078-9959-1be0d259342a req-017426b5-1bf3-4c2b-a660-c998100f4002 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.812633] env[61995]: DEBUG nova.network.neutron [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 846.031525] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794627, 'name': Rename_Task, 'duration_secs': 0.143025} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.031975] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 846.032333] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-325272ed-93c4-474d-9663-786ab03c2c21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.042146] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 846.042146] env[61995]: value = "task-794628" [ 846.042146] env[61995]: _type = "Task" [ 846.042146] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.053663] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.079902] env[61995]: DEBUG nova.compute.utils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.083522] env[61995]: DEBUG nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.083698] env[61995]: DEBUG nova.network.neutron [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 846.092668] env[61995]: DEBUG nova.compute.manager [req-805dceba-8acd-43d4-93bb-0356c53b5731 req-3364c9b6-30c1-4341-a55e-943128a45607 service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Received event network-vif-deleted-9116dcbf-d617-485e-afde-c82e908d5c15 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.092863] env[61995]: INFO nova.compute.manager [req-805dceba-8acd-43d4-93bb-0356c53b5731 req-3364c9b6-30c1-4341-a55e-943128a45607 service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Neutron deleted interface 9116dcbf-d617-485e-afde-c82e908d5c15; detaching it from the instance and deleting it from the info cache [ 846.093157] env[61995]: DEBUG nova.network.neutron [req-805dceba-8acd-43d4-93bb-0356c53b5731 req-3364c9b6-30c1-4341-a55e-943128a45607 service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.141141] env[61995]: DEBUG nova.policy [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2388a3bd1587413285ff9f9ec07739b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9301d14211a1464eb740d9bb745b631a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 846.153781] env[61995]: DEBUG nova.network.neutron [-] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.182731] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-080df9c2-4ccc-4be3-9597-b2df6b8bb199 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.193551] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b0f133-116a-4bd9-b206-d53881cf1779 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.234309] env[61995]: DEBUG nova.compute.manager [req-5fbdde08-31bd-4078-9959-1be0d259342a req-017426b5-1bf3-4c2b-a660-c998100f4002 service nova] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Detach interface failed, port_id=9f82eee3-593f-46f6-a461-4c7aec380a4e, reason: Instance d3f629c5-96f4-4208-be0d-cbf9810ecc6b could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 846.494767] env[61995]: DEBUG nova.network.neutron [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Successfully created port: 3ecd14b2-d6b9-4216-936e-edae029037af {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.541687] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970bac12-d72b-49a2-a7b0-f54fdb1429b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.556640] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477c7279-352d-4370-b95a-a74634edefd3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.559687] env[61995]: DEBUG oslo_vmware.api [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794628, 'name': PowerOnVM_Task, 'duration_secs': 0.445395} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.559945] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 846.560169] env[61995]: DEBUG nova.compute.manager [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.561194] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f165a0-c4ec-425d-986c-33520e872564 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.588064] env[61995]: DEBUG nova.network.neutron [-] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.592947] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d889d1-dad5-4cc5-af4a-cd16dd568d95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.595665] env[61995]: DEBUG nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.603079] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dead87eb-9d05-4e5c-bb94-9909ccb453d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.611824] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183459ca-62fe-458f-8fd2-4f06403ab850 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.620162] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a51dc3-16ab-48d6-9fda-9f99ede2520e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.639924] env[61995]: DEBUG nova.compute.provider_tree [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.652019] env[61995]: DEBUG nova.compute.manager [req-805dceba-8acd-43d4-93bb-0356c53b5731 req-3364c9b6-30c1-4341-a55e-943128a45607 service nova] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Detach interface failed, port_id=9116dcbf-d617-485e-afde-c82e908d5c15, reason: Instance eef1cf32-e71e-4696-8170-1023e3a5518d could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 846.653238] env[61995]: DEBUG nova.scheduler.client.report [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.660021] env[61995]: DEBUG nova.network.neutron [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.660021] env[61995]: INFO nova.compute.manager [-] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Took 1.68 seconds to deallocate network for instance. [ 847.092545] env[61995]: INFO nova.compute.manager [-] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Took 1.58 seconds to deallocate network for instance. [ 847.114535] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.160079] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.585s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.162818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Releasing lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.162818] env[61995]: DEBUG nova.compute.manager [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Inject network info {{(pid=61995) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 847.163070] env[61995]: DEBUG nova.compute.manager [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] network_info to inject: |[{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 847.169162] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Reconfiguring VM instance to set the machine id {{(pid=61995) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 847.169717] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 23.507s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.173291] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.173550] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1669d785-86d3-4a97-8710-0cee243ec4a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.192861] env[61995]: DEBUG oslo_vmware.api [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 847.192861] env[61995]: value = "task-794629" [ 847.192861] env[61995]: _type = "Task" [ 847.192861] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.206034] env[61995]: DEBUG oslo_vmware.api [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794629, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.208803] env[61995]: INFO nova.scheduler.client.report [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted allocations for instance 823d2169-2bcb-4d5a-9677-cc4ac49a7e84 [ 847.510262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.510404] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.510618] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.510801] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.510969] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.515535] env[61995]: INFO nova.compute.manager [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Terminating instance [ 847.519820] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "refresh_cache-966a72d8-a51a-44f9-8184-8108f8cc3ce6" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.520581] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquired lock "refresh_cache-966a72d8-a51a-44f9-8184-8108f8cc3ce6" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.520847] env[61995]: DEBUG nova.network.neutron [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.600828] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.613228] env[61995]: DEBUG nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.640544] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.640790] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.641048] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.641147] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.641302] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.641455] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.642287] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.645617] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.645843] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.646123] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.646330] env[61995]: DEBUG nova.virt.hardware [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.647200] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702df88c-e304-4a09-9360-3c1df16a705e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.658427] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca39e0cb-f8fa-4bd8-9487-119bfeeaf5de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.703880] env[61995]: DEBUG oslo_vmware.api [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794629, 'name': ReconfigVM_Task, 'duration_secs': 0.144812} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.704086] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4615-255b-46f4-bb69-8b4b80759ec0 tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Reconfigured VM instance to set the machine id {{(pid=61995) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 847.714984] env[61995]: DEBUG nova.compute.manager [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Received event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 847.714984] env[61995]: DEBUG nova.compute.manager [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing instance network info cache due to event network-changed-d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 847.714984] env[61995]: DEBUG oslo_concurrency.lockutils [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] Acquiring lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.714984] env[61995]: DEBUG oslo_concurrency.lockutils [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] Acquired lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.714984] env[61995]: DEBUG nova.network.neutron [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Refreshing network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.720800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85f863ec-8ec0-4ed1-8032-0bd0a7d5ec24 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "823d2169-2bcb-4d5a-9677-cc4ac49a7e84" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.187s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.861371] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.862275] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.862764] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.863042] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.863297] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.867213] env[61995]: INFO nova.compute.manager [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Terminating instance [ 847.872135] env[61995]: DEBUG nova.compute.manager [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 847.872378] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 847.873219] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a536814-7d5d-4c28-a032-85ad1677ac63 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.882154] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 847.882416] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20ee3ac1-3f78-466c-8816-ed3d74bebbff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.890088] env[61995]: DEBUG oslo_vmware.api [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 847.890088] env[61995]: value = "task-794630" [ 847.890088] env[61995]: _type = "Task" [ 847.890088] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.898747] env[61995]: DEBUG oslo_vmware.api [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.220373] env[61995]: DEBUG nova.network.neutron [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 848.224452] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e53d10a3-1d16-498d-b4d8-abe7205410ec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 848.224717] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d9c53806-68c0-4872-a262-36bc05573674 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.224717] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b1bd98d9-bd0f-4abd-a188-e5267ada4852 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.224826] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance c4f744f9-36f3-4913-9ced-f0db93d3b4df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.224935] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance fcdc7c33-8c46-4f4d-bb39-e6f943256cb0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.225059] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance f9c1424f-6096-481c-aa91-acc25bfd8f6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.225194] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance eef1cf32-e71e-4696-8170-1023e3a5518d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 848.225311] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.225420] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance dc7077ac-d3fd-4e84-867a-d86328b32f5d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.225529] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b40b1866-ced3-40ef-9ab7-30d58b75f288 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.225637] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.225757] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 0348098c-fd4c-41f8-a246-b9ba87a75674 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 848.225865] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 4edbf65d-eaca-47e1-b6bf-03ccb908b52e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.226025] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 7524c912-4d73-462a-949f-3b4c85874094 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 848.226118] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance bb31ef64-478f-4f39-9ee2-fcfabe8235ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.226587] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance af7163e8-4e77-4336-80e9-0623e29ed8c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.226728] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 451e226c-5fb7-4042-a706-3e58b0606546 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 848.226850] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 966a72d8-a51a-44f9-8184-8108f8cc3ce6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.226975] env[61995]: WARNING nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d3f629c5-96f4-4208-be0d-cbf9810ecc6b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 848.227103] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 2b40609a-0826-4a16-af28-c8926af6b646 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 848.311972] env[61995]: DEBUG nova.compute.manager [req-f88339e0-a9a2-41e7-a883-a1d2587c6dab req-d137037a-696a-4272-b8ea-b6bebd576d63 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Received event network-vif-plugged-3ecd14b2-d6b9-4216-936e-edae029037af {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.312338] env[61995]: DEBUG oslo_concurrency.lockutils [req-f88339e0-a9a2-41e7-a883-a1d2587c6dab req-d137037a-696a-4272-b8ea-b6bebd576d63 service nova] Acquiring lock "2b40609a-0826-4a16-af28-c8926af6b646-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.312830] env[61995]: DEBUG oslo_concurrency.lockutils [req-f88339e0-a9a2-41e7-a883-a1d2587c6dab req-d137037a-696a-4272-b8ea-b6bebd576d63 service nova] Lock "2b40609a-0826-4a16-af28-c8926af6b646-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.313034] env[61995]: DEBUG oslo_concurrency.lockutils [req-f88339e0-a9a2-41e7-a883-a1d2587c6dab req-d137037a-696a-4272-b8ea-b6bebd576d63 service nova] Lock "2b40609a-0826-4a16-af28-c8926af6b646-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.313216] env[61995]: DEBUG nova.compute.manager [req-f88339e0-a9a2-41e7-a883-a1d2587c6dab req-d137037a-696a-4272-b8ea-b6bebd576d63 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] No waiting events found dispatching network-vif-plugged-3ecd14b2-d6b9-4216-936e-edae029037af {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.313390] env[61995]: WARNING nova.compute.manager [req-f88339e0-a9a2-41e7-a883-a1d2587c6dab req-d137037a-696a-4272-b8ea-b6bebd576d63 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Received unexpected event network-vif-plugged-3ecd14b2-d6b9-4216-936e-edae029037af for instance with vm_state building and task_state spawning. [ 848.364424] env[61995]: DEBUG nova.network.neutron [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.375315] env[61995]: DEBUG nova.network.neutron [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Successfully updated port: 3ecd14b2-d6b9-4216-936e-edae029037af {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.401831] env[61995]: DEBUG oslo_vmware.api [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794630, 'name': PowerOffVM_Task, 'duration_secs': 0.324574} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.402176] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 848.402314] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 848.402575] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1f016b4-5b7c-44d5-90ce-8fb08f68c700 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.507618] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 848.509010] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 848.509010] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Deleting the datastore file [datastore2] f9c1424f-6096-481c-aa91-acc25bfd8f6e {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.509684] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d352b0f-13c7-42a0-ad31-8d28fa14471d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.515908] env[61995]: DEBUG oslo_vmware.api [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for the task: (returnval){ [ 848.515908] env[61995]: value = "task-794632" [ 848.515908] env[61995]: _type = "Task" [ 848.515908] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.525249] env[61995]: DEBUG oslo_vmware.api [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.731985] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 848.788128] env[61995]: DEBUG nova.network.neutron [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updated VIF entry in instance network info cache for port d044777d-78bf-4b5b-a0dd-3b2dec017fbc. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 848.788537] env[61995]: DEBUG nova.network.neutron [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [{"id": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "address": "fa:16:3e:ed:57:40", "network": {"id": "48c0e237-2e25-4514-98b6-b29dd702e626", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1859010663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1b27447bed34a35911f7585608107b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd044777d-78", "ovs_interfaceid": "d044777d-78bf-4b5b-a0dd-3b2dec017fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.867584] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Releasing lock "refresh_cache-966a72d8-a51a-44f9-8184-8108f8cc3ce6" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.868083] env[61995]: DEBUG nova.compute.manager [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.868299] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 848.869214] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2beae8c2-2e37-4555-afe9-eac2818fcdfb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.879283] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.879283] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.879283] env[61995]: DEBUG nova.network.neutron [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 848.879610] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 848.882607] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba097a38-f0fe-4a49-a31f-384136f632b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.886240] env[61995]: DEBUG oslo_vmware.api [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 848.886240] env[61995]: value = "task-794633" [ 848.886240] env[61995]: _type = "Task" [ 848.886240] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.897495] env[61995]: DEBUG oslo_vmware.api [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.026365] env[61995]: DEBUG oslo_vmware.api [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.236751] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 229d966c-b70d-4a51-a176-2e88488d5c4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 849.295860] env[61995]: DEBUG oslo_concurrency.lockutils [req-b8e52ff8-71a7-4cc9-8b53-ed4cee078507 req-90d77a58-9015-4e33-b186-70902d1d9a2f service nova] Releasing lock "refresh_cache-f9c1424f-6096-481c-aa91-acc25bfd8f6e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.395599] env[61995]: DEBUG oslo_vmware.api [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794633, 'name': PowerOffVM_Task, 'duration_secs': 0.248131} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.395808] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.395976] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.396273] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08096a1a-4f21-4210-9be2-66a9e1e4861f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.407946] env[61995]: DEBUG nova.network.neutron [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 849.425222] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.425453] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.425685] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Deleting the datastore file [datastore1] 966a72d8-a51a-44f9-8184-8108f8cc3ce6 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.425978] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31800883-5f18-450b-af22-3074de260624 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.434815] env[61995]: DEBUG oslo_vmware.api [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for the task: (returnval){ [ 849.434815] env[61995]: value = "task-794635" [ 849.434815] env[61995]: _type = "Task" [ 849.434815] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.445699] env[61995]: DEBUG oslo_vmware.api [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.529851] env[61995]: DEBUG oslo_vmware.api [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Task: {'id': task-794632, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.544707} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.530152] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.530347] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 849.530529] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 849.530707] env[61995]: INFO nova.compute.manager [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Took 1.66 seconds to destroy the instance on the hypervisor. [ 849.530966] env[61995]: DEBUG oslo.service.loopingcall [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.531184] env[61995]: DEBUG nova.compute.manager [-] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.531283] env[61995]: DEBUG nova.network.neutron [-] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.575504] env[61995]: DEBUG nova.network.neutron [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance_info_cache with network_info: [{"id": "3ecd14b2-d6b9-4216-936e-edae029037af", "address": "fa:16:3e:2b:38:33", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ecd14b2-d6", "ovs_interfaceid": "3ecd14b2-d6b9-4216-936e-edae029037af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.590909] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.591172] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.740141] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance cfcf1a5a-97db-4095-b6a6-044242ee8a05 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 849.945600] env[61995]: DEBUG oslo_vmware.api [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Task: {'id': task-794635, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152314} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.945714] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.945915] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 849.946204] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 849.946471] env[61995]: INFO nova.compute.manager [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Took 1.08 seconds to destroy the instance on the hypervisor. [ 849.946805] env[61995]: DEBUG oslo.service.loopingcall [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.947083] env[61995]: DEBUG nova.compute.manager [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.947228] env[61995]: DEBUG nova.network.neutron [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.975575] env[61995]: DEBUG nova.network.neutron [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.082021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.082021] env[61995]: DEBUG nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Instance network_info: |[{"id": "3ecd14b2-d6b9-4216-936e-edae029037af", "address": "fa:16:3e:2b:38:33", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ecd14b2-d6", "ovs_interfaceid": "3ecd14b2-d6b9-4216-936e-edae029037af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.082021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:38:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ecd14b2-d6b9-4216-936e-edae029037af', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.094327] env[61995]: DEBUG oslo.service.loopingcall [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.095183] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 850.095570] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69f82d12-1071-4f3d-ba96-9b903dccd135 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.114451] env[61995]: DEBUG nova.compute.utils [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.122784] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.122784] env[61995]: value = "task-794636" [ 850.122784] env[61995]: _type = "Task" [ 850.122784] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.134574] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794636, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.243141] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ed952baf-f1ef-4ec0-8a15-928287190456 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 850.480019] env[61995]: DEBUG nova.network.neutron [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.505776] env[61995]: DEBUG nova.network.neutron [-] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.600175] env[61995]: DEBUG nova.compute.manager [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Received event network-changed-3ecd14b2-d6b9-4216-936e-edae029037af {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.600175] env[61995]: DEBUG nova.compute.manager [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Refreshing instance network info cache due to event network-changed-3ecd14b2-d6b9-4216-936e-edae029037af. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 850.600175] env[61995]: DEBUG oslo_concurrency.lockutils [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] Acquiring lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.600175] env[61995]: DEBUG oslo_concurrency.lockutils [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] Acquired lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.600175] env[61995]: DEBUG nova.network.neutron [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Refreshing network info cache for port 3ecd14b2-d6b9-4216-936e-edae029037af {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 850.617638] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.026s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.633219] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794636, 'name': CreateVM_Task, 'duration_secs': 0.389896} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.633438] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 850.634222] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.634422] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.634795] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.635152] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d8cb237-edc6-4a63-a130-f4f4d0e0af7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.642606] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 850.642606] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f2c651-b9cd-75c7-dea0-204a65e4aa86" [ 850.642606] env[61995]: _type = "Task" [ 850.642606] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.650607] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f2c651-b9cd-75c7-dea0-204a65e4aa86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.746758] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 4b49e7b1-7fd0-4e59-af79-5c0898967c35 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 850.747625] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 850.747625] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 850.981751] env[61995]: INFO nova.compute.manager [-] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Took 1.03 seconds to deallocate network for instance. [ 851.012942] env[61995]: INFO nova.compute.manager [-] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Took 1.48 seconds to deallocate network for instance. [ 851.089050] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee42e6c-36ef-4f75-9169-c7511d9fbe01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.096728] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b85349c-5591-4d5e-9461-62ba776f0522 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.136935] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7776454-2d02-4a36-a1c4-15bc5f5ab475 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.145702] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf0c4c3-68df-430a-857d-c310ffb8a298 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.155110] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f2c651-b9cd-75c7-dea0-204a65e4aa86, 'name': SearchDatastore_Task, 'duration_secs': 0.009908} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.165717] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.166009] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.166444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.166614] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.166843] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.167626] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.168811] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e085862a-228b-43b3-92b1-f3d57fd7bffb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.176909] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.177192] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 851.177960] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13a23b0b-b30d-4268-9a40-737b986a1dbf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.186047] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 851.186047] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52002284-4e4e-c03b-5568-535fc3ef442c" [ 851.186047] env[61995]: _type = "Task" [ 851.186047] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.194730] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52002284-4e4e-c03b-5568-535fc3ef442c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.404197] env[61995]: DEBUG nova.network.neutron [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updated VIF entry in instance network info cache for port 3ecd14b2-d6b9-4216-936e-edae029037af. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 851.404514] env[61995]: DEBUG nova.network.neutron [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance_info_cache with network_info: [{"id": "3ecd14b2-d6b9-4216-936e-edae029037af", "address": "fa:16:3e:2b:38:33", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ecd14b2-d6", "ovs_interfaceid": "3ecd14b2-d6b9-4216-936e-edae029037af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.487675] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.526900] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.672399] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.696740] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52002284-4e4e-c03b-5568-535fc3ef442c, 'name': SearchDatastore_Task, 'duration_secs': 0.008612} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.697514] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eebd4de1-cba6-4a77-9121-0156989ad699 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.702647] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 851.702647] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521d9a59-5e4a-6450-a990-0092f596f396" [ 851.702647] env[61995]: _type = "Task" [ 851.702647] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.711531] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.711763] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.711987] env[61995]: INFO nova.compute.manager [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Attaching volume 34345e8c-85a0-4c01-85e5-bc6ec70b8cc9 to /dev/sdb [ 851.713884] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521d9a59-5e4a-6450-a990-0092f596f396, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.745148] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7a6893-52be-40f7-b4f5-d9e0aa9fd665 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.751306] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be51526b-e84e-4515-9643-4d5d7e16ea53 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.766063] env[61995]: DEBUG nova.virt.block_device [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Updating existing volume attachment record: 5713e6e7-20fe-4c73-8141-530d5d70e6b3 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 851.906912] env[61995]: DEBUG oslo_concurrency.lockutils [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] Releasing lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.907272] env[61995]: DEBUG nova.compute.manager [req-4fb509bd-3ad6-44f4-8594-6986251d80bf req-37a72077-ed2e-492e-bc98-ee0a4eb9ba41 service nova] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Received event network-vif-deleted-d044777d-78bf-4b5b-a0dd-3b2dec017fbc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.178129] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 852.178425] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.009s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.178740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.878s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.180529] env[61995]: INFO nova.compute.claims [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.184163] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.184163] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Cleaning up deleted instances {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 852.215559] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521d9a59-5e4a-6450-a990-0092f596f396, 'name': SearchDatastore_Task, 'duration_secs': 0.009618} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.216311] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.216588] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2b40609a-0826-4a16-af28-c8926af6b646/2b40609a-0826-4a16-af28-c8926af6b646.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 852.216872] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d0b52a71-309b-4790-8420-7b007a680958 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.224625] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 852.224625] env[61995]: value = "task-794644" [ 852.224625] env[61995]: _type = "Task" [ 852.224625] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.236044] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.694103] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] There are 22 instances to clean {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 852.694446] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 823d2169-2bcb-4d5a-9677-cc4ac49a7e84] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 852.735125] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476382} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.736029] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 2b40609a-0826-4a16-af28-c8926af6b646/2b40609a-0826-4a16-af28-c8926af6b646.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 852.736345] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.736844] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1d5932d-a52f-4e5b-8290-65d5475e10d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.744045] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 852.744045] env[61995]: value = "task-794645" [ 852.744045] env[61995]: _type = "Task" [ 852.744045] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.753871] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794645, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.200087] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 0d34066f-5b8f-4bac-9b83-67d78987c4b3] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 853.255552] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794645, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.307962} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.255807] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.258772] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0581476-9dec-40f7-92d0-b69e647014bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.281842] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 2b40609a-0826-4a16-af28-c8926af6b646/2b40609a-0826-4a16-af28-c8926af6b646.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.284619] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67b3eba4-0c94-41ef-a4c0-110987316e98 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.304032] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 853.304032] env[61995]: value = "task-794646" [ 853.304032] env[61995]: _type = "Task" [ 853.304032] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.313399] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794646, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.670329] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c813937f-5d0a-4880-b26c-ebffecea47a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.677817] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3e2bf0-0c70-4766-93f2-a158d319f593 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.707093] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d056dadd-492d-4a4d-abc2-fc5f113f865b] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 853.709607] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c796ff-0f73-48c3-881c-f059a3a0c3c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.717802] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff0dc1d-be8a-4265-b1b1-5138642342b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.731532] env[61995]: DEBUG nova.compute.provider_tree [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.822854] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794646, 'name': ReconfigVM_Task, 'duration_secs': 0.323787} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.823385] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 2b40609a-0826-4a16-af28-c8926af6b646/2b40609a-0826-4a16-af28-c8926af6b646.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.824258] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee3384d8-5181-4184-892d-86d74e42e6ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.836181] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 853.836181] env[61995]: value = "task-794647" [ 853.836181] env[61995]: _type = "Task" [ 853.836181] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.850042] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794647, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.213017] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: f6db353f-f7b5-4ba4-8f70-f3b65281da93] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 854.235796] env[61995]: DEBUG nova.scheduler.client.report [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.346452] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794647, 'name': Rename_Task, 'duration_secs': 0.167892} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.346757] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 854.347061] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6488f711-74fb-43fb-84d3-2d029f4e4c58 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.353443] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 854.353443] env[61995]: value = "task-794650" [ 854.353443] env[61995]: _type = "Task" [ 854.353443] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.362790] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.717014] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: c51184b6-d2c7-49ff-b061-ab0f9a69cdbf] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 854.741835] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.741835] env[61995]: DEBUG nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.745015] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.401s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.745873] env[61995]: INFO nova.compute.claims [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.864063] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794650, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.901582] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.901826] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.219928] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 2da17a43-2260-4a82-9ed3-eafd25bb9a6a] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 855.252402] env[61995]: DEBUG nova.compute.utils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.253871] env[61995]: DEBUG nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.254053] env[61995]: DEBUG nova.network.neutron [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 855.301819] env[61995]: DEBUG nova.policy [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce97dbfe04a349738a0e5ee9cf177338', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a269227b55044d38154a50fbc722723', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 855.366735] env[61995]: DEBUG oslo_vmware.api [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794650, 'name': PowerOnVM_Task, 'duration_secs': 0.531296} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.367033] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 855.367252] env[61995]: INFO nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Took 7.75 seconds to spawn the instance on the hypervisor. [ 855.367432] env[61995]: DEBUG nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.368265] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc330ef-2f58-4582-b2cc-12818c7066b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.404583] env[61995]: DEBUG nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.581418] env[61995]: DEBUG nova.network.neutron [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Successfully created port: b25c9bc9-03d6-4bcd-a3e6-8469128cea27 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.723096] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 4da426a1-7af9-4e4e-82ed-c45687a19b86] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 855.756876] env[61995]: DEBUG nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.889878] env[61995]: INFO nova.compute.manager [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Took 35.39 seconds to build instance. [ 855.925052] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.084476] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f8cdc1-b999-407b-bf1a-89a5c5f674a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.092833] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1bbec9-db21-486d-b50c-e278613e1f1b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.123078] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da05721-7707-4b88-a56f-0f88dc794774 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.130777] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc3fbd0-f3aa-4a3d-82c6-c857164a65ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.144713] env[61995]: DEBUG nova.compute.provider_tree [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.227806] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 0d7238b3-d7a3-4a35-ba33-bf2e7e7a9afa] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 856.319555] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 856.320161] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185322', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'name': 'volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc7077ac-d3fd-4e84-867a-d86328b32f5d', 'attached_at': '', 'detached_at': '', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'serial': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 856.321237] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e17cd6d-9c7d-440f-bd9c-4072a2d329b6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.340337] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc44d15c-0b97-4b18-b209-6a9183d6164b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.368586] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9/volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.368876] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27a2e5e6-3be6-4fad-aa2e-0bfc566eb921 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.387831] env[61995]: DEBUG oslo_vmware.api [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 856.387831] env[61995]: value = "task-794652" [ 856.387831] env[61995]: _type = "Task" [ 856.387831] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.391703] env[61995]: DEBUG oslo_concurrency.lockutils [None req-46c1911b-3850-4117-acf3-62a5d4bd7a7a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.917s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.395906] env[61995]: DEBUG oslo_vmware.api [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794652, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.647643] env[61995]: DEBUG nova.scheduler.client.report [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.734487] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ecb72668-caa4-4117-9f8d-e0c6e4831f3d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 856.766071] env[61995]: DEBUG nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.794531] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.794895] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.794964] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.795147] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.795304] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.795457] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.795669] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.795842] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.796043] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.796225] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.796404] env[61995]: DEBUG nova.virt.hardware [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.797571] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f528c5d-671d-4e16-aec3-86f68c96f524 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.805957] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc45e69a-ad62-4312-9d24-13bafa861a87 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.902606] env[61995]: DEBUG oslo_vmware.api [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794652, 'name': ReconfigVM_Task, 'duration_secs': 0.336874} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.902913] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Reconfigured VM instance instance-00000034 to attach disk [datastore2] volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9/volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.907949] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39989773-90b1-48f2-a406-08717126e20e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.923062] env[61995]: DEBUG oslo_vmware.api [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 856.923062] env[61995]: value = "task-794653" [ 856.923062] env[61995]: _type = "Task" [ 856.923062] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.931342] env[61995]: DEBUG oslo_vmware.api [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794653, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.041930] env[61995]: DEBUG nova.compute.manager [req-f6b0f748-e6f0-4378-8e05-85e8af3b1e43 req-8a48a424-5348-4f01-b2c8-9e2961e0d531 service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Received event network-vif-plugged-b25c9bc9-03d6-4bcd-a3e6-8469128cea27 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 857.042310] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6b0f748-e6f0-4378-8e05-85e8af3b1e43 req-8a48a424-5348-4f01-b2c8-9e2961e0d531 service nova] Acquiring lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.042566] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6b0f748-e6f0-4378-8e05-85e8af3b1e43 req-8a48a424-5348-4f01-b2c8-9e2961e0d531 service nova] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.042752] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6b0f748-e6f0-4378-8e05-85e8af3b1e43 req-8a48a424-5348-4f01-b2c8-9e2961e0d531 service nova] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.042941] env[61995]: DEBUG nova.compute.manager [req-f6b0f748-e6f0-4378-8e05-85e8af3b1e43 req-8a48a424-5348-4f01-b2c8-9e2961e0d531 service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] No waiting events found dispatching network-vif-plugged-b25c9bc9-03d6-4bcd-a3e6-8469128cea27 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 857.043187] env[61995]: WARNING nova.compute.manager [req-f6b0f748-e6f0-4378-8e05-85e8af3b1e43 req-8a48a424-5348-4f01-b2c8-9e2961e0d531 service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Received unexpected event network-vif-plugged-b25c9bc9-03d6-4bcd-a3e6-8469128cea27 for instance with vm_state building and task_state spawning. [ 857.152284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.152835] env[61995]: DEBUG nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.155890] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.701s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.156106] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.158069] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.059s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.159436] env[61995]: INFO nova.compute.claims [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.174018] env[61995]: DEBUG nova.network.neutron [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Successfully updated port: b25c9bc9-03d6-4bcd-a3e6-8469128cea27 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.191698] env[61995]: INFO nova.scheduler.client.report [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Deleted allocations for instance 0348098c-fd4c-41f8-a246-b9ba87a75674 [ 857.237566] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 55555deb-9883-483b-a901-42b1a66c60d6] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 857.433683] env[61995]: DEBUG oslo_vmware.api [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794653, 'name': ReconfigVM_Task, 'duration_secs': 0.153094} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.434087] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185322', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'name': 'volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc7077ac-d3fd-4e84-867a-d86328b32f5d', 'attached_at': '', 'detached_at': '', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'serial': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 857.664402] env[61995]: DEBUG nova.compute.utils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.668783] env[61995]: DEBUG nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.668865] env[61995]: DEBUG nova.network.neutron [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 857.677962] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.678134] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.678745] env[61995]: DEBUG nova.network.neutron [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 857.698130] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1a1c35ef-dd9e-49cb-8b00-97f0719fafe5 tempest-ServersTestMultiNic-803055499 tempest-ServersTestMultiNic-803055499-project-member] Lock "0348098c-fd4c-41f8-a246-b9ba87a75674" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.672s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.711981] env[61995]: DEBUG nova.policy [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4fa18e6b6fe742909fa15a846d3019e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd23f8abd8f14ec392fbfb7fd5bc64f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 857.741399] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a04fbe7e-e53e-43ef-9a6f-ba3895ce7597] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 858.013054] env[61995]: DEBUG nova.network.neutron [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Successfully created port: cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.053012] env[61995]: DEBUG nova.compute.manager [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 858.172520] env[61995]: DEBUG nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.216707] env[61995]: DEBUG nova.network.neutron [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.244229] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: c350194c-ed3e-4996-871a-70b4cb99c7db] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 858.403699] env[61995]: DEBUG nova.network.neutron [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Updating instance_info_cache with network_info: [{"id": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "address": "fa:16:3e:33:8d:a1", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25c9bc9-03", "ovs_interfaceid": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.484614] env[61995]: DEBUG nova.objects.instance [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'flavor' on Instance uuid dc7077ac-d3fd-4e84-867a-d86328b32f5d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.572152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.576442] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0fada8-ddd5-48b5-ab74-19c299b4fbab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.584026] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6eb79e-be06-4f13-b018-bec6a38eeab5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.620841] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119d3a3e-45e3-405d-89ac-d6bea74ca30a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.629528] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868360b0-ff06-43a4-98ee-de243cbf58fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.643366] env[61995]: DEBUG nova.compute.provider_tree [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.746423] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 19fd0358-32e4-434f-b0b3-c8cc6c07ba51] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 858.908942] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.909329] env[61995]: DEBUG nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Instance network_info: |[{"id": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "address": "fa:16:3e:33:8d:a1", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25c9bc9-03", "ovs_interfaceid": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.909766] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:8d:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b25c9bc9-03d6-4bcd-a3e6-8469128cea27', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.919780] env[61995]: DEBUG oslo.service.loopingcall [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.920020] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 858.921063] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38204a88-e2ac-48bd-b284-4d6e0f6c0339 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.941603] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.941603] env[61995]: value = "task-794655" [ 858.941603] env[61995]: _type = "Task" [ 858.941603] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.954707] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794655, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.990913] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9c11a4b-509f-4027-ae33-4602b8609699 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.278s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.076207] env[61995]: DEBUG nova.compute.manager [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Received event network-changed-b25c9bc9-03d6-4bcd-a3e6-8469128cea27 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 859.076207] env[61995]: DEBUG nova.compute.manager [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Refreshing instance network info cache due to event network-changed-b25c9bc9-03d6-4bcd-a3e6-8469128cea27. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 859.076207] env[61995]: DEBUG oslo_concurrency.lockutils [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] Acquiring lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.076207] env[61995]: DEBUG oslo_concurrency.lockutils [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] Acquired lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.076207] env[61995]: DEBUG nova.network.neutron [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Refreshing network info cache for port b25c9bc9-03d6-4bcd-a3e6-8469128cea27 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.146840] env[61995]: DEBUG nova.scheduler.client.report [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.184298] env[61995]: DEBUG nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.216896] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.217358] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.217721] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.218054] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.218929] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.218929] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.218929] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.218929] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.219299] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.221570] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.221570] env[61995]: DEBUG nova.virt.hardware [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.221690] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e1842c-5986-4e86-bea4-2de6e25fc2cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.230542] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586614d7-21d5-452b-bce1-a26698b1f899 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.250124] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 45fdae21-bec2-438a-8f2a-c23b0e428057] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 859.453084] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794655, 'name': CreateVM_Task, 'duration_secs': 0.438755} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.453319] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.454150] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.454333] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.454668] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.455440] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35d3465d-b8b6-460d-89d8-3f17aa49ea8a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.461688] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 859.461688] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52abde5b-5282-4acd-f9c2-42161704a992" [ 859.461688] env[61995]: _type = "Task" [ 859.461688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.471855] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52abde5b-5282-4acd-f9c2-42161704a992, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.653666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.654274] env[61995]: DEBUG nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.656952] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.140s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.657175] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.659288] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.203s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.660858] env[61995]: INFO nova.compute.claims [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.701390] env[61995]: INFO nova.scheduler.client.report [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleted allocations for instance 7524c912-4d73-462a-949f-3b4c85874094 [ 859.755578] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 24db5fad-52fd-4689-a8aa-2ae4b0a06710] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 859.893183] env[61995]: DEBUG nova.network.neutron [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Updated VIF entry in instance network info cache for port b25c9bc9-03d6-4bcd-a3e6-8469128cea27. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.893183] env[61995]: DEBUG nova.network.neutron [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Updating instance_info_cache with network_info: [{"id": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "address": "fa:16:3e:33:8d:a1", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25c9bc9-03", "ovs_interfaceid": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.982021] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52abde5b-5282-4acd-f9c2-42161704a992, 'name': SearchDatastore_Task, 'duration_secs': 0.009525} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.982021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.982021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.982021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.982021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.982021] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.982021] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-084e38d8-4b50-4bf4-b028-e9cc2691013d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.991769] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.992307] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.993318] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d62f09f-5893-449b-9e17-dd03edb98b14 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.000522] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 860.000522] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52fd3d5b-d5bb-572b-a258-c3650a49dd27" [ 860.000522] env[61995]: _type = "Task" [ 860.000522] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.014118] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fd3d5b-d5bb-572b-a258-c3650a49dd27, 'name': SearchDatastore_Task, 'duration_secs': 0.008396} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.014981] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8d43d4d-1105-4086-9284-8d6175e1d93d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.022022] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 860.022022] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520866a5-9df2-393c-e175-a34a415670a8" [ 860.022022] env[61995]: _type = "Task" [ 860.022022] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.029510] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520866a5-9df2-393c-e175-a34a415670a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.167712] env[61995]: DEBUG nova.compute.utils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.173183] env[61995]: DEBUG nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Not allocating networking since 'none' was specified. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 860.215269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b42eb91e-45f0-4930-82d6-1147f83ec4c9 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "7524c912-4d73-462a-949f-3b4c85874094" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.891s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.223422] env[61995]: DEBUG nova.network.neutron [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Successfully updated port: cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.260648] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 1c4eb398-c4e9-4588-ba48-2805e3fee750] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 860.394781] env[61995]: DEBUG oslo_concurrency.lockutils [req-ea441c27-3251-424f-96a1-5aa41bc93e7a req-b2b4bb5e-fc76-4d0f-801d-2753fd4d3cab service nova] Releasing lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.534249] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520866a5-9df2-393c-e175-a34a415670a8, 'name': SearchDatastore_Task, 'duration_secs': 0.008642} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.537020] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.537020] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400/5480efdf-fd46-4a26-b6a5-b6c3dd4c4400.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.537020] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5119f98-129d-459a-8b36-eee4305c0c82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.547196] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 860.547196] env[61995]: value = "task-794657" [ 860.547196] env[61995]: _type = "Task" [ 860.547196] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.557363] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794657, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.639400] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.640013] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.673058] env[61995]: DEBUG nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.728487] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.729094] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.729094] env[61995]: DEBUG nova.network.neutron [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.771851] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 08489e01-2ce3-45a1-8f87-2a4c89230aa2] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 861.060608] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794657, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.140940] env[61995]: DEBUG nova.compute.manager [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Received event network-vif-plugged-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.141568] env[61995]: DEBUG oslo_concurrency.lockutils [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] Acquiring lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.141568] env[61995]: DEBUG oslo_concurrency.lockutils [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.141568] env[61995]: DEBUG oslo_concurrency.lockutils [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.142159] env[61995]: DEBUG nova.compute.manager [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] No waiting events found dispatching network-vif-plugged-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.142159] env[61995]: WARNING nova.compute.manager [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Received unexpected event network-vif-plugged-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 for instance with vm_state building and task_state spawning. [ 861.142400] env[61995]: DEBUG nova.compute.manager [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Received event network-changed-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.142400] env[61995]: DEBUG nova.compute.manager [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Refreshing instance network info cache due to event network-changed-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 861.142564] env[61995]: DEBUG oslo_concurrency.lockutils [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] Acquiring lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.143576] env[61995]: DEBUG nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.191261] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09c37d5-6185-449a-82ff-e07345029c73 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.201173] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21442560-6123-4e41-884b-843056906318 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.238823] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257efee6-798b-412a-b7b4-e5304932c9ce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.248395] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048a5673-72b4-4274-8444-b9627bf44782 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.262268] env[61995]: DEBUG nova.compute.provider_tree [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.275507] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: f47e5bbb-f4a8-4686-8195-8348d2ffb08d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 861.304559] env[61995]: DEBUG nova.network.neutron [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.516859] env[61995]: DEBUG nova.network.neutron [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.557064] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565537} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.557626] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400/5480efdf-fd46-4a26-b6a5-b6c3dd4c4400.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.557626] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.557917] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-997a2fc3-f458-4eee-9d4c-1b082dcbd522 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.568637] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 861.568637] env[61995]: value = "task-794658" [ 861.568637] env[61995]: _type = "Task" [ 861.568637] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.580182] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794658, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.666870] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.686499] env[61995]: DEBUG nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.711330] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.712030] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.712030] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.712030] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.712184] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.712258] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.712478] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.712640] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.712814] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.712983] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.713183] env[61995]: DEBUG nova.virt.hardware [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.714052] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a0ded7-8fe7-4fed-b9dd-edbaa393f4b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.722653] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6356c602-c99d-4cb7-a0cb-897fcb8f88cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.736758] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.742428] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Creating folder: Project (3ba52f931cdb41728f250a458cda943e). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 861.742773] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5e5ee7e-78a9-40da-8911-cd94dccd83da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.755478] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Created folder: Project (3ba52f931cdb41728f250a458cda943e) in parent group-v185203. [ 861.755763] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Creating folder: Instances. Parent ref: group-v185325. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 861.756068] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-553877e1-5290-4fc9-b014-db879dd91ccd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.766374] env[61995]: DEBUG nova.scheduler.client.report [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.770520] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Created folder: Instances in parent group-v185325. [ 861.770836] env[61995]: DEBUG oslo.service.loopingcall [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.771022] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 861.771301] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a821e4c6-fb9f-447a-bbbf-b3ff3cb14318 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.786398] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e78f8718-e0be-4b2a-ab70-9565ce01ab9d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 861.795646] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.795646] env[61995]: value = "task-794661" [ 861.795646] env[61995]: _type = "Task" [ 861.795646] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.803922] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794661, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.020707] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.020707] env[61995]: DEBUG nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Instance network_info: |[{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.020978] env[61995]: DEBUG oslo_concurrency.lockutils [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] Acquired lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.021177] env[61995]: DEBUG nova.network.neutron [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Refreshing network info cache for port cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 862.022449] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:68:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7894814c-6be3-4b80-a08e-4a771bc05dd1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.031790] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating folder: Project (cd23f8abd8f14ec392fbfb7fd5bc64f9). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.035706] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05a21830-eba9-4f8b-b40d-98e812a5702d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.051470] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Created folder: Project (cd23f8abd8f14ec392fbfb7fd5bc64f9) in parent group-v185203. [ 862.051732] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating folder: Instances. Parent ref: group-v185328. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.052009] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b5d9187-a9a6-4d69-a5e7-dc9e59faa2da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.062627] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Created folder: Instances in parent group-v185328. [ 862.062913] env[61995]: DEBUG oslo.service.loopingcall [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.063151] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 862.063386] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-150bb4ae-2284-4306-878e-b38e79b60d97 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.092654] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794658, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088692} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.093855] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.094142] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.094142] env[61995]: value = "task-794664" [ 862.094142] env[61995]: _type = "Task" [ 862.094142] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.097789] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef843862-1eaf-4f9b-bbea-972f5efad49b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.130619] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400/5480efdf-fd46-4a26-b6a5-b6c3dd4c4400.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.130982] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f59cd9e-7af3-469b-aab7-f4dabaca102d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.151489] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 862.151489] env[61995]: value = "task-794665" [ 862.151489] env[61995]: _type = "Task" [ 862.151489] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.163129] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794665, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.271803] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.272453] env[61995]: DEBUG nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.275010] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.988s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.275222] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.277183] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.778s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.277381] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.279772] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.070s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.281704] env[61995]: INFO nova.compute.claims [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.296850] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d5a54c46-1a45-4bd5-8694-0defc757c455] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 862.308298] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794661, 'name': CreateVM_Task, 'duration_secs': 0.387116} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.308519] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.308984] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.309140] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.309483] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.309763] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-602d7921-f84c-4d5e-a374-8e01de924675 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.315538] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 862.315538] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523fdb87-8bd2-1b9e-f43a-b33d9d2be502" [ 862.315538] env[61995]: _type = "Task" [ 862.315538] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.325035] env[61995]: INFO nova.scheduler.client.report [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Deleted allocations for instance e53d10a3-1d16-498d-b4d8-abe7205410ec [ 862.327104] env[61995]: INFO nova.scheduler.client.report [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Deleted allocations for instance 451e226c-5fb7-4042-a706-3e58b0606546 [ 862.331784] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523fdb87-8bd2-1b9e-f43a-b33d9d2be502, 'name': SearchDatastore_Task, 'duration_secs': 0.010498} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.335926] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.336251] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.336467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.336622] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.336804] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.338711] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b22ce68d-0a4a-450d-a56f-42a339441af9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.350523] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.350661] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 862.351583] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bef3b6f9-e544-4994-9445-0e49e39a37b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.357875] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 862.357875] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]524db9ad-8379-ae4e-d031-93d131091e55" [ 862.357875] env[61995]: _type = "Task" [ 862.357875] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.363168] env[61995]: DEBUG nova.network.neutron [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updated VIF entry in instance network info cache for port cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 862.363571] env[61995]: DEBUG nova.network.neutron [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.369296] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524db9ad-8379-ae4e-d031-93d131091e55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.414367] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.414720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.414977] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.415293] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.415537] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.418203] env[61995]: INFO nova.compute.manager [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Terminating instance [ 862.420504] env[61995]: DEBUG nova.compute.manager [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.420719] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.421566] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a932d7-a548-4e93-b4a1-d8a3c06dff62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.428727] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.428989] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bcd2ba68-f3f0-4cca-ae3f-8a87dc0393db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.435318] env[61995]: DEBUG oslo_vmware.api [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 862.435318] env[61995]: value = "task-794666" [ 862.435318] env[61995]: _type = "Task" [ 862.435318] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.445568] env[61995]: DEBUG oslo_vmware.api [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794666, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.477090] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "af7163e8-4e77-4336-80e9-0623e29ed8c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.477445] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.477762] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "af7163e8-4e77-4336-80e9-0623e29ed8c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.478056] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.478284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.480945] env[61995]: INFO nova.compute.manager [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Terminating instance [ 862.483652] env[61995]: DEBUG nova.compute.manager [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.483864] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.484766] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26245180-200f-4e07-97e8-6d460e8b1ef7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.493185] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.493845] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78debc78-e351-4fe1-a5a8-2a8c1e4dffa0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.501850] env[61995]: DEBUG oslo_vmware.api [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 862.501850] env[61995]: value = "task-794667" [ 862.501850] env[61995]: _type = "Task" [ 862.501850] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.511570] env[61995]: DEBUG oslo_vmware.api [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794667, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.612739] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794664, 'name': CreateVM_Task, 'duration_secs': 0.48397} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.612997] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.613753] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.613972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.614366] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.614686] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-687aab13-0f56-4f9e-962a-2a97c52a6f99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.620692] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 862.620692] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52bb697f-2950-28b0-2b0b-3be27da443fc" [ 862.620692] env[61995]: _type = "Task" [ 862.620692] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.629522] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52bb697f-2950-28b0-2b0b-3be27da443fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.663415] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794665, 'name': ReconfigVM_Task, 'duration_secs': 0.296139} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.663776] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400/5480efdf-fd46-4a26-b6a5-b6c3dd4c4400.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.664423] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3cbacdda-64eb-4fe2-837a-6d2410981df8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.671141] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 862.671141] env[61995]: value = "task-794668" [ 862.671141] env[61995]: _type = "Task" [ 862.671141] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.682361] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794668, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.790257] env[61995]: DEBUG nova.compute.utils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.791123] env[61995]: DEBUG nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.791292] env[61995]: DEBUG nova.network.neutron [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 862.802348] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 06496d6a-324d-41d4-b3e5-52e3d93636bb] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 862.847086] env[61995]: DEBUG oslo_concurrency.lockutils [None req-30f23f05-16a0-4440-918d-30554f3f325f tempest-FloatingIPsAssociationTestJSON-1227433622 tempest-FloatingIPsAssociationTestJSON-1227433622-project-member] Lock "e53d10a3-1d16-498d-b4d8-abe7205410ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.602s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.847086] env[61995]: DEBUG oslo_concurrency.lockutils [None req-440ae491-78e3-4b8e-8e3a-e4c60f678f7e tempest-ServerGroupTestJSON-1427464096 tempest-ServerGroupTestJSON-1427464096-project-member] Lock "451e226c-5fb7-4042-a706-3e58b0606546" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.065s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.864568] env[61995]: DEBUG nova.policy [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caf9bbbd848d4d7ba53ba3bd553943e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2018aae65cf4bf98b5ac82519201cf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 862.866231] env[61995]: DEBUG oslo_concurrency.lockutils [req-1274c65b-c581-4e3d-bbb0-2d79412ef62c req-f4b45dfd-0cfc-4170-9cb7-fcad8c5768ba service nova] Releasing lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.871832] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524db9ad-8379-ae4e-d031-93d131091e55, 'name': SearchDatastore_Task, 'duration_secs': 0.010145} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.874728] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7451a049-9d93-48c7-bc57-7f780e50ccfd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.882252] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 862.882252] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522080e2-bf7b-0610-e197-6706c26f5f23" [ 862.882252] env[61995]: _type = "Task" [ 862.882252] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.890793] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522080e2-bf7b-0610-e197-6706c26f5f23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.947747] env[61995]: DEBUG oslo_vmware.api [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794666, 'name': PowerOffVM_Task, 'duration_secs': 0.187015} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.947992] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 862.948221] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 862.948471] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2a5007e-59df-422a-9f14-e106f303551e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.012584] env[61995]: DEBUG oslo_vmware.api [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794667, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.052047] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.052299] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.052489] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleting the datastore file [datastore1] bb31ef64-478f-4f39-9ee2-fcfabe8235ec {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.052756] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff655162-1c26-4555-b7b7-10c1cc717400 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.059898] env[61995]: DEBUG oslo_vmware.api [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 863.059898] env[61995]: value = "task-794670" [ 863.059898] env[61995]: _type = "Task" [ 863.059898] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.068695] env[61995]: DEBUG oslo_vmware.api [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.132737] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52bb697f-2950-28b0-2b0b-3be27da443fc, 'name': SearchDatastore_Task, 'duration_secs': 0.010331} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.132737] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.133409] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.133409] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.182317] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794668, 'name': Rename_Task, 'duration_secs': 0.146763} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.182720] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 863.183012] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6634378-41f3-4209-bc8e-28e9b74d8b9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.193478] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 863.193478] env[61995]: value = "task-794671" [ 863.193478] env[61995]: _type = "Task" [ 863.193478] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.207269] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.296450] env[61995]: DEBUG nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.305008] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 2eeecb76-ff2d-4a9d-a86f-9e2b8e1fc6cc] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 863.397402] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522080e2-bf7b-0610-e197-6706c26f5f23, 'name': SearchDatastore_Task, 'duration_secs': 0.010082} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.397706] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.398083] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] cfcf1a5a-97db-4095-b6a6-044242ee8a05/cfcf1a5a-97db-4095-b6a6-044242ee8a05.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 863.398317] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.399018] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.399018] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6695b96c-438c-4406-aa5c-144df863c687 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.403542] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63bb177d-18ca-4307-a781-df775e5a29be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.411945] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 863.411945] env[61995]: value = "task-794672" [ 863.411945] env[61995]: _type = "Task" [ 863.411945] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.421457] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.421457] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 863.423620] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ae92bc3-7315-4936-83cf-fe8bf3b08eb8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.434176] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794672, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.437205] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 863.437205] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52362936-bc52-c96e-6e29-e2f3aad47f69" [ 863.437205] env[61995]: _type = "Task" [ 863.437205] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.448405] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52362936-bc52-c96e-6e29-e2f3aad47f69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.493434] env[61995]: DEBUG nova.network.neutron [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Successfully created port: 8df0df4d-9ab4-4900-a107-027ae7ffe387 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.514131] env[61995]: DEBUG oslo_vmware.api [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794667, 'name': PowerOffVM_Task, 'duration_secs': 0.792074} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.514131] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.514131] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.514131] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e00363d9-2414-4812-8a40-17e0dc331487 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.573814] env[61995]: DEBUG oslo_vmware.api [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145241} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.578218] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.578485] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 863.578689] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 863.578875] env[61995]: INFO nova.compute.manager [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Took 1.16 seconds to destroy the instance on the hypervisor. [ 863.579317] env[61995]: DEBUG oslo.service.loopingcall [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.579657] env[61995]: DEBUG nova.compute.manager [-] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.579747] env[61995]: DEBUG nova.network.neutron [-] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 863.586438] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.586576] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.586687] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleting the datastore file [datastore1] af7163e8-4e77-4336-80e9-0623e29ed8c0 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.587296] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb6776b8-c9bd-4a7f-b937-29df03c2a685 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.594845] env[61995]: DEBUG oslo_vmware.api [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for the task: (returnval){ [ 863.594845] env[61995]: value = "task-794674" [ 863.594845] env[61995]: _type = "Task" [ 863.594845] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.606831] env[61995]: DEBUG oslo_vmware.api [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.716247] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794671, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.753611] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3d4dc7-2707-4032-a83d-f3c029e41b22 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.764042] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b776d502-0cbc-4bca-9bfd-e7e99da41fe9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.799300] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c1a69e-eefe-4c62-a64f-906e01bb0a11 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.811702] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 863.811991] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Cleaning up deleted instances with incomplete migration {{(pid=61995) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 863.814364] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3e4cef-abb7-4015-a0cf-a09ab94a4ae7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.835282] env[61995]: DEBUG nova.compute.provider_tree [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.927632] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794672, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.954387] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52362936-bc52-c96e-6e29-e2f3aad47f69, 'name': SearchDatastore_Task, 'duration_secs': 0.014797} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.955503] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0b6e615-7cde-4b14-843b-95054efe8641 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.964081] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 863.964081] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e4e5b4-7534-85aa-0cf0-4b49d902ad85" [ 863.964081] env[61995]: _type = "Task" [ 863.964081] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.975751] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e4e5b4-7534-85aa-0cf0-4b49d902ad85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.112490] env[61995]: DEBUG oslo_vmware.api [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Task: {'id': task-794674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343955} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.112490] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.112490] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 864.112490] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.112490] env[61995]: INFO nova.compute.manager [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Took 1.63 seconds to destroy the instance on the hypervisor. [ 864.112490] env[61995]: DEBUG oslo.service.loopingcall [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.112490] env[61995]: DEBUG nova.compute.manager [-] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.112490] env[61995]: DEBUG nova.network.neutron [-] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.211221] env[61995]: DEBUG oslo_vmware.api [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794671, 'name': PowerOnVM_Task, 'duration_secs': 0.767045} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.212944] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 864.213199] env[61995]: INFO nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Took 7.45 seconds to spawn the instance on the hypervisor. [ 864.213398] env[61995]: DEBUG nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 864.214822] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b429a38-1e37-40cd-98af-2480d693ec65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.313344] env[61995]: DEBUG nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.322408] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 864.339833] env[61995]: DEBUG nova.scheduler.client.report [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.357023] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.357023] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.357023] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.357023] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.357023] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.358044] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.358459] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.358778] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.359226] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.359519] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.359809] env[61995]: DEBUG nova.virt.hardware [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.362030] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5b69bb-484c-46a8-89c5-8b3a19b61561 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.374111] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e9dc25-56b0-4df7-a920-e0f5a363862c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.415253] env[61995]: DEBUG nova.compute.manager [req-5908bdff-38c5-4fea-8d18-1014ec822dd6 req-b6a0bf98-9320-4a8d-996e-3db5426650aa service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Received event network-vif-deleted-73460573-883d-4a1e-86f5-2c257ff21ea0 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.415496] env[61995]: INFO nova.compute.manager [req-5908bdff-38c5-4fea-8d18-1014ec822dd6 req-b6a0bf98-9320-4a8d-996e-3db5426650aa service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Neutron deleted interface 73460573-883d-4a1e-86f5-2c257ff21ea0; detaching it from the instance and deleting it from the info cache [ 864.415701] env[61995]: DEBUG nova.network.neutron [req-5908bdff-38c5-4fea-8d18-1014ec822dd6 req-b6a0bf98-9320-4a8d-996e-3db5426650aa service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.426774] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794672, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.749923} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.427104] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] cfcf1a5a-97db-4095-b6a6-044242ee8a05/cfcf1a5a-97db-4095-b6a6-044242ee8a05.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 864.427334] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.427589] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81cb4a94-6a1f-49ba-a315-c4f537918be8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.439524] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 864.439524] env[61995]: value = "task-794676" [ 864.439524] env[61995]: _type = "Task" [ 864.439524] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.450333] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794676, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.477308] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e4e5b4-7534-85aa-0cf0-4b49d902ad85, 'name': SearchDatastore_Task, 'duration_secs': 0.056588} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.477308] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.477525] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 229d966c-b70d-4a51-a176-2e88488d5c4b/229d966c-b70d-4a51-a176-2e88488d5c4b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 864.477631] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0517705e-fae0-4495-bc03-b0bbfc6491c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.484605] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 864.484605] env[61995]: value = "task-794677" [ 864.484605] env[61995]: _type = "Task" [ 864.484605] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.494882] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794677, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.714873] env[61995]: DEBUG nova.network.neutron [-] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.736744] env[61995]: INFO nova.compute.manager [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Took 40.47 seconds to build instance. [ 864.845108] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.845837] env[61995]: DEBUG nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.852114] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.738s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.852114] env[61995]: DEBUG nova.objects.instance [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 864.923528] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1ce9b57-f9a5-4038-9b9e-af5cf5d45d58 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.934423] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80627692-d940-43a7-bcd5-560cb81c9f90 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.962347] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794676, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105868} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.962347] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.963154] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe3e03d-bea3-42a2-9b1f-8fb0e88ac844 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.978705] env[61995]: DEBUG nova.compute.manager [req-5908bdff-38c5-4fea-8d18-1014ec822dd6 req-b6a0bf98-9320-4a8d-996e-3db5426650aa service nova] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Detach interface failed, port_id=73460573-883d-4a1e-86f5-2c257ff21ea0, reason: Instance bb31ef64-478f-4f39-9ee2-fcfabe8235ec could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 865.000612] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] cfcf1a5a-97db-4095-b6a6-044242ee8a05/cfcf1a5a-97db-4095-b6a6-044242ee8a05.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.003912] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbb8e28d-8ec8-4508-bf89-2b67319e4969 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.025536] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794677, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.027226] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 865.027226] env[61995]: value = "task-794678" [ 865.027226] env[61995]: _type = "Task" [ 865.027226] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.036307] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794678, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.128416] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] Acquiring lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.128627] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] Acquired lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.128811] env[61995]: DEBUG nova.network.neutron [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 865.153519] env[61995]: DEBUG nova.network.neutron [-] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.217844] env[61995]: INFO nova.compute.manager [-] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Took 1.64 seconds to deallocate network for instance. [ 865.240637] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5ac4046e-1c6a-40f0-9194-f6b93eaa393b tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.975s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.358282] env[61995]: DEBUG nova.compute.utils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.364202] env[61995]: DEBUG nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 865.368040] env[61995]: DEBUG nova.network.neutron [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 865.486900] env[61995]: DEBUG nova.policy [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c135f484cfe0401f83d1fcf16debf8b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a7fbb0e40964f349256c7529874ca8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 865.502524] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794677, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.959149} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.502817] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 229d966c-b70d-4a51-a176-2e88488d5c4b/229d966c-b70d-4a51-a176-2e88488d5c4b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 865.503058] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.503332] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13035f02-f838-47c8-82e0-49b844d0f78d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.513082] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 865.513082] env[61995]: value = "task-794679" [ 865.513082] env[61995]: _type = "Task" [ 865.513082] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.528581] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.539530] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794678, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.657128] env[61995]: INFO nova.compute.manager [-] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Took 1.55 seconds to deallocate network for instance. [ 865.726059] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.794616] env[61995]: DEBUG nova.network.neutron [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Successfully updated port: 8df0df4d-9ab4-4900-a107-027ae7ffe387 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.873468] env[61995]: DEBUG nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.880334] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d88c0b5-e27e-47d4-87b8-ae4428e7aa35 tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.028s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.883080] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.710s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.883415] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.886277] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.286s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.886744] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.888736] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.401s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.888998] env[61995]: DEBUG nova.objects.instance [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lazy-loading 'resources' on Instance uuid 966a72d8-a51a-44f9-8184-8108f8cc3ce6 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.923131] env[61995]: INFO nova.scheduler.client.report [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleted allocations for instance d3f629c5-96f4-4208-be0d-cbf9810ecc6b [ 865.925678] env[61995]: INFO nova.scheduler.client.report [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Deleted allocations for instance eef1cf32-e71e-4696-8170-1023e3a5518d [ 865.972220] env[61995]: DEBUG nova.compute.manager [req-9b470160-ffa8-4958-90fd-e0537cd931e8 req-fbee3722-d0ac-4ba6-b890-c1d69290fba8 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Received event network-vif-plugged-8df0df4d-9ab4-4900-a107-027ae7ffe387 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.972949] env[61995]: DEBUG oslo_concurrency.lockutils [req-9b470160-ffa8-4958-90fd-e0537cd931e8 req-fbee3722-d0ac-4ba6-b890-c1d69290fba8 service nova] Acquiring lock "ed952baf-f1ef-4ec0-8a15-928287190456-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.973844] env[61995]: DEBUG oslo_concurrency.lockutils [req-9b470160-ffa8-4958-90fd-e0537cd931e8 req-fbee3722-d0ac-4ba6-b890-c1d69290fba8 service nova] Lock "ed952baf-f1ef-4ec0-8a15-928287190456-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.973844] env[61995]: DEBUG oslo_concurrency.lockutils [req-9b470160-ffa8-4958-90fd-e0537cd931e8 req-fbee3722-d0ac-4ba6-b890-c1d69290fba8 service nova] Lock "ed952baf-f1ef-4ec0-8a15-928287190456-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.973844] env[61995]: DEBUG nova.compute.manager [req-9b470160-ffa8-4958-90fd-e0537cd931e8 req-fbee3722-d0ac-4ba6-b890-c1d69290fba8 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] No waiting events found dispatching network-vif-plugged-8df0df4d-9ab4-4900-a107-027ae7ffe387 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.974421] env[61995]: WARNING nova.compute.manager [req-9b470160-ffa8-4958-90fd-e0537cd931e8 req-fbee3722-d0ac-4ba6-b890-c1d69290fba8 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Received unexpected event network-vif-plugged-8df0df4d-9ab4-4900-a107-027ae7ffe387 for instance with vm_state building and task_state spawning. [ 866.026321] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071641} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.026658] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.027538] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5848db68-f84b-4803-b47f-48c717377c24 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.038567] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794678, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.059496] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 229d966c-b70d-4a51-a176-2e88488d5c4b/229d966c-b70d-4a51-a176-2e88488d5c4b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.060167] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c400397e-6c2b-4f7b-920e-6d58d10bb5d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.090504] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 866.090504] env[61995]: value = "task-794680" [ 866.090504] env[61995]: _type = "Task" [ 866.090504] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.107156] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794680, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.166044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.166579] env[61995]: DEBUG nova.network.neutron [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Successfully created port: 9eb488f7-7474-4b91-9b01-2f4378f5daca {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.264072] env[61995]: DEBUG nova.network.neutron [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Updating instance_info_cache with network_info: [{"id": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "address": "fa:16:3e:33:8d:a1", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25c9bc9-03", "ovs_interfaceid": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.296987] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "refresh_cache-ed952baf-f1ef-4ec0-8a15-928287190456" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.297274] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "refresh_cache-ed952baf-f1ef-4ec0-8a15-928287190456" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.297319] env[61995]: DEBUG nova.network.neutron [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.442289] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1e1c543c-1da4-413b-bc14-27afb840d4e4 tempest-ServersTestManualDisk-1774981459 tempest-ServersTestManualDisk-1774981459-project-member] Lock "eef1cf32-e71e-4696-8170-1023e3a5518d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.139s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.443392] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6b6d82f-8789-4c36-91bc-762681ba61ec tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "d3f629c5-96f4-4208-be0d-cbf9810ecc6b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.645s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.547649] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794678, 'name': ReconfigVM_Task, 'duration_secs': 1.086204} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.548459] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Reconfigured VM instance instance-00000041 to attach disk [datastore2] cfcf1a5a-97db-4095-b6a6-044242ee8a05/cfcf1a5a-97db-4095-b6a6-044242ee8a05.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.551022] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7446cfa0-24ff-4a64-adc1-227463215df9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.562017] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 866.562017] env[61995]: value = "task-794682" [ 866.562017] env[61995]: _type = "Task" [ 866.562017] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.572663] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794682, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.600890] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794680, 'name': ReconfigVM_Task, 'duration_secs': 0.354048} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.604097] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 229d966c-b70d-4a51-a176-2e88488d5c4b/229d966c-b70d-4a51-a176-2e88488d5c4b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.605101] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dade767f-b305-44fc-8351-d1d1c4cd7067 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.616043] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 866.616043] env[61995]: value = "task-794683" [ 866.616043] env[61995]: _type = "Task" [ 866.616043] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.628109] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794683, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.724936] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13919cc5-55ce-4991-a1f8-104c9a8b73b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.734771] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27efbdb7-4127-444f-958f-b847893a94dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.768142] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] Releasing lock "refresh_cache-5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.768142] env[61995]: DEBUG nova.compute.manager [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Inject network info {{(pid=61995) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 866.768405] env[61995]: DEBUG nova.compute.manager [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] network_info to inject: |[{"id": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "address": "fa:16:3e:33:8d:a1", "network": {"id": "cfbb9937-4dfe-4488-afc8-2dc49390c159", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1409973401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a269227b55044d38154a50fbc722723", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25c9bc9-03", "ovs_interfaceid": "b25c9bc9-03d6-4bcd-a3e6-8469128cea27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 866.773811] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Reconfiguring VM instance to set the machine id {{(pid=61995) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 866.774513] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddcaacf4-53cf-4c59-8df2-7a64ad6bb426 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.786651] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7833d6be-0145-4cae-b51c-f81ffe02928b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.795465] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c01244a-b8f5-462c-b004-b31237b9cc99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.800681] env[61995]: DEBUG oslo_vmware.api [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] Waiting for the task: (returnval){ [ 866.800681] env[61995]: value = "task-794684" [ 866.800681] env[61995]: _type = "Task" [ 866.800681] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.814843] env[61995]: DEBUG nova.compute.provider_tree [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.822564] env[61995]: DEBUG oslo_vmware.api [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] Task: {'id': task-794684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.839381] env[61995]: DEBUG nova.compute.manager [req-5590c53c-6c9b-41db-b4a2-1b3de760d3a8 req-d42a73b2-1bef-422a-bd1d-ab51a1f5a657 service nova] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Received event network-vif-deleted-6910a7c3-0d3f-4941-8135-018a19f8d5ef {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.872286] env[61995]: DEBUG nova.network.neutron [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 866.888223] env[61995]: DEBUG nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.916242] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:58:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2ec0e7a2-ad31-4b6b-b0cd-3d774a987d95',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-892104472',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.916387] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.916793] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.916793] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.916915] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.916967] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.918185] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.918432] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.918635] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.918836] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.919055] env[61995]: DEBUG nova.virt.hardware [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.919993] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28b1dbe-d40f-4649-b7a4-535bdaf0a9b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.931034] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aceb9559-237c-46ff-b4db-b03c12f267c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.071470] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794682, 'name': Rename_Task, 'duration_secs': 0.22134} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.071808] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.072078] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99c835bd-fc4a-4233-a298-9999b5e4cd0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.080039] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 867.080039] env[61995]: value = "task-794685" [ 867.080039] env[61995]: _type = "Task" [ 867.080039] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.088871] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794685, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.129108] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794683, 'name': Rename_Task, 'duration_secs': 0.171012} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.129644] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.130146] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed95a155-4b8d-4d4e-8162-d790379f7900 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.138985] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 867.138985] env[61995]: value = "task-794686" [ 867.138985] env[61995]: _type = "Task" [ 867.138985] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.147656] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794686, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.315574] env[61995]: DEBUG oslo_vmware.api [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] Task: {'id': task-794684, 'name': ReconfigVM_Task, 'duration_secs': 0.254233} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.315959] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a35b8918-905e-493c-8db6-419c1ba54148 tempest-ServersAdminTestJSON-234259859 tempest-ServersAdminTestJSON-234259859-project-admin] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Reconfigured VM instance to set the machine id {{(pid=61995) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 867.321020] env[61995]: DEBUG nova.scheduler.client.report [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.476030] env[61995]: DEBUG nova.network.neutron [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Updating instance_info_cache with network_info: [{"id": "8df0df4d-9ab4-4900-a107-027ae7ffe387", "address": "fa:16:3e:04:a6:07", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8df0df4d-9a", "ovs_interfaceid": "8df0df4d-9ab4-4900-a107-027ae7ffe387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.596154] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794685, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.652235] env[61995]: DEBUG oslo_vmware.api [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794686, 'name': PowerOnVM_Task, 'duration_secs': 0.482567} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.652235] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 867.652235] env[61995]: INFO nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Took 8.47 seconds to spawn the instance on the hypervisor. [ 867.652516] env[61995]: DEBUG nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.653329] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68025218-d557-473f-b35b-b5ff0e86226e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.823808] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.935s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.826437] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.299s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.826688] env[61995]: DEBUG nova.objects.instance [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lazy-loading 'resources' on Instance uuid f9c1424f-6096-481c-aa91-acc25bfd8f6e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.853616] env[61995]: INFO nova.scheduler.client.report [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Deleted allocations for instance 966a72d8-a51a-44f9-8184-8108f8cc3ce6 [ 867.983025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "refresh_cache-ed952baf-f1ef-4ec0-8a15-928287190456" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.983025] env[61995]: DEBUG nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Instance network_info: |[{"id": "8df0df4d-9ab4-4900-a107-027ae7ffe387", "address": "fa:16:3e:04:a6:07", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8df0df4d-9a", "ovs_interfaceid": "8df0df4d-9ab4-4900-a107-027ae7ffe387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.983025] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:a6:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8df0df4d-9ab4-4900-a107-027ae7ffe387', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.994847] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating folder: Project (b2018aae65cf4bf98b5ac82519201cf4). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 867.995423] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3046a04-3749-4565-a9b4-a926875ebe1b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.011995] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created folder: Project (b2018aae65cf4bf98b5ac82519201cf4) in parent group-v185203. [ 868.013457] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating folder: Instances. Parent ref: group-v185331. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 868.013457] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c666a89-2cb5-4c64-88c6-8d25d7a82568 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.026088] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created folder: Instances in parent group-v185331. [ 868.026088] env[61995]: DEBUG oslo.service.loopingcall [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.026088] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 868.026088] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-323dedf2-506b-4d15-8f68-8fd328b40ef3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.049225] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 868.049225] env[61995]: value = "task-794689" [ 868.049225] env[61995]: _type = "Task" [ 868.049225] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.058433] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794689, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.072516] env[61995]: DEBUG nova.compute.manager [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Received event network-changed-8df0df4d-9ab4-4900-a107-027ae7ffe387 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.072588] env[61995]: DEBUG nova.compute.manager [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Refreshing instance network info cache due to event network-changed-8df0df4d-9ab4-4900-a107-027ae7ffe387. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 868.074377] env[61995]: DEBUG oslo_concurrency.lockutils [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] Acquiring lock "refresh_cache-ed952baf-f1ef-4ec0-8a15-928287190456" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.079097] env[61995]: DEBUG oslo_concurrency.lockutils [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] Acquired lock "refresh_cache-ed952baf-f1ef-4ec0-8a15-928287190456" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.079097] env[61995]: DEBUG nova.network.neutron [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Refreshing network info cache for port 8df0df4d-9ab4-4900-a107-027ae7ffe387 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 868.098181] env[61995]: DEBUG oslo_vmware.api [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794685, 'name': PowerOnVM_Task, 'duration_secs': 0.89704} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.098369] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 868.098533] env[61995]: INFO nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Took 6.41 seconds to spawn the instance on the hypervisor. [ 868.098715] env[61995]: DEBUG nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.099692] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200b2c22-8ffc-4a95-9f40-8d28cf5bfdde {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.177674] env[61995]: INFO nova.compute.manager [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Took 41.86 seconds to build instance. [ 868.364040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1b775d8b-718e-4d9e-8c5b-181fab1f032e tempest-ServerShowV257Test-818965244 tempest-ServerShowV257Test-818965244-project-member] Lock "966a72d8-a51a-44f9-8184-8108f8cc3ce6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.853s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.530297] env[61995]: DEBUG nova.network.neutron [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Successfully updated port: 9eb488f7-7474-4b91-9b01-2f4378f5daca {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.565545] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794689, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.620535] env[61995]: INFO nova.compute.manager [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Took 32.54 seconds to build instance. [ 868.685502] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f75d06ed-32db-4461-9259-7ce41eb1d145 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.391s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.703022] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f256cc-a82e-4a7a-90a8-0c0573cf4cbc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.716100] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bfa990-6a4a-440f-a94d-40b88f55666f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.754117] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2396a562-f257-45c1-bffb-8c2349454fb4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.763667] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47f5dbd-daa3-4d13-b971-e7a65d10c6dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.784711] env[61995]: DEBUG nova.compute.provider_tree [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.945753] env[61995]: DEBUG nova.compute.manager [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Received event network-vif-plugged-9eb488f7-7474-4b91-9b01-2f4378f5daca {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.945975] env[61995]: DEBUG oslo_concurrency.lockutils [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] Acquiring lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.946224] env[61995]: DEBUG oslo_concurrency.lockutils [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.946393] env[61995]: DEBUG oslo_concurrency.lockutils [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.946558] env[61995]: DEBUG nova.compute.manager [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] No waiting events found dispatching network-vif-plugged-9eb488f7-7474-4b91-9b01-2f4378f5daca {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 868.946723] env[61995]: WARNING nova.compute.manager [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Received unexpected event network-vif-plugged-9eb488f7-7474-4b91-9b01-2f4378f5daca for instance with vm_state building and task_state spawning. [ 868.946879] env[61995]: DEBUG nova.compute.manager [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Received event network-changed-9eb488f7-7474-4b91-9b01-2f4378f5daca {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.948780] env[61995]: DEBUG nova.compute.manager [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Refreshing instance network info cache due to event network-changed-9eb488f7-7474-4b91-9b01-2f4378f5daca. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 868.949089] env[61995]: DEBUG oslo_concurrency.lockutils [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] Acquiring lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.949243] env[61995]: DEBUG oslo_concurrency.lockutils [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] Acquired lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.949413] env[61995]: DEBUG nova.network.neutron [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Refreshing network info cache for port 9eb488f7-7474-4b91-9b01-2f4378f5daca {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 869.034388] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.055140] env[61995]: DEBUG nova.network.neutron [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Updated VIF entry in instance network info cache for port 8df0df4d-9ab4-4900-a107-027ae7ffe387. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 869.055511] env[61995]: DEBUG nova.network.neutron [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Updating instance_info_cache with network_info: [{"id": "8df0df4d-9ab4-4900-a107-027ae7ffe387", "address": "fa:16:3e:04:a6:07", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8df0df4d-9a", "ovs_interfaceid": "8df0df4d-9ab4-4900-a107-027ae7ffe387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.072663] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794689, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.123473] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9b46741-0061-48d5-bf46-a9adcb147dbf tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.066s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.294799] env[61995]: DEBUG nova.scheduler.client.report [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.541971] env[61995]: DEBUG nova.network.neutron [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 869.558829] env[61995]: DEBUG oslo_concurrency.lockutils [req-457f5556-b8a0-4e87-895e-246bd18a32f3 req-0c053f1c-1caf-47d5-beef-1744b1f3b924 service nova] Releasing lock "refresh_cache-ed952baf-f1ef-4ec0-8a15-928287190456" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.564599] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794689, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.767244] env[61995]: INFO nova.compute.manager [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Rebuilding instance [ 869.803147] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.807842] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.883s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.809985] env[61995]: INFO nova.compute.claims [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.834895] env[61995]: DEBUG nova.compute.manager [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.836187] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65b4e3d-ab1c-4365-8317-086d269b45db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.839777] env[61995]: INFO nova.scheduler.client.report [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Deleted allocations for instance f9c1424f-6096-481c-aa91-acc25bfd8f6e [ 869.942692] env[61995]: DEBUG nova.network.neutron [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.018640] env[61995]: DEBUG nova.compute.manager [None req-1a44d52e-dc47-480c-8a53-743171c60d17 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.019636] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2deea36-369d-4c6d-ba05-81d21d1e906e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.068148] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794689, 'name': CreateVM_Task, 'duration_secs': 1.621815} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.068148] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 870.069057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.069057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.069458] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.070097] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-929c9275-6850-4c60-8174-af77596b6d6b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.078951] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 870.078951] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526ae2ef-b697-02c9-2aa3-6b9553db6cf4" [ 870.078951] env[61995]: _type = "Task" [ 870.078951] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.088693] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526ae2ef-b697-02c9-2aa3-6b9553db6cf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.357564] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 870.359328] env[61995]: DEBUG oslo_concurrency.lockutils [None req-97ddfde1-18e3-415d-8034-1dc1414cbcde tempest-AttachInterfacesUnderV243Test-308775099 tempest-AttachInterfacesUnderV243Test-308775099-project-member] Lock "f9c1424f-6096-481c-aa91-acc25bfd8f6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.496s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.359328] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba8965f9-0d2f-485c-9148-6a587462b050 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.370188] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 870.370188] env[61995]: value = "task-794691" [ 870.370188] env[61995]: _type = "Task" [ 870.370188] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.383000] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794691, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.449863] env[61995]: DEBUG oslo_concurrency.lockutils [req-72f7eddb-6241-436b-b769-4aea836f6ace req-e798e142-081c-436c-9f71-61b79c5f20a9 service nova] Releasing lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.449863] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.449863] env[61995]: DEBUG nova.network.neutron [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 870.533434] env[61995]: INFO nova.compute.manager [None req-1a44d52e-dc47-480c-8a53-743171c60d17 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] instance snapshotting [ 870.534414] env[61995]: DEBUG nova.objects.instance [None req-1a44d52e-dc47-480c-8a53-743171c60d17 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lazy-loading 'flavor' on Instance uuid cfcf1a5a-97db-4095-b6a6-044242ee8a05 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.591106] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526ae2ef-b697-02c9-2aa3-6b9553db6cf4, 'name': SearchDatastore_Task, 'duration_secs': 0.012955} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.591843] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.592147] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.592441] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.592751] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.592998] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.593331] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f73c75fa-cb09-43a8-8242-05c2b370dfa1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.609529] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.609529] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 870.609817] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fe0a7f5-27aa-41d0-90cb-4465ec2b38ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.616879] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 870.616879] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52daad9c-f902-ec40-1d27-554fc2218c6c" [ 870.616879] env[61995]: _type = "Task" [ 870.616879] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.628128] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52daad9c-f902-ec40-1d27-554fc2218c6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.763537] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.764066] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.765439] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.766066] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.766066] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.769883] env[61995]: INFO nova.compute.manager [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Terminating instance [ 870.773789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "refresh_cache-cfcf1a5a-97db-4095-b6a6-044242ee8a05" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.773939] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquired lock "refresh_cache-cfcf1a5a-97db-4095-b6a6-044242ee8a05" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.774126] env[61995]: DEBUG nova.network.neutron [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 870.881138] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794691, 'name': PowerOffVM_Task, 'duration_secs': 0.198875} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.883972] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 870.884260] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 870.885757] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2673dfd-e8cf-40a5-85d7-7b78655cb9b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.893588] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 870.894136] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7265acde-72c7-45a2-bc18-71c5c6803ed9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.006371] env[61995]: DEBUG nova.network.neutron [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 871.050963] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95b6fcc-3c9e-4c87-975a-0db40c5be15f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.088564] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de77adfc-fe92-4f3d-afc4-57843e43ad5d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.103960] env[61995]: DEBUG nova.compute.manager [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Received event network-changed-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.106093] env[61995]: DEBUG nova.compute.manager [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Refreshing instance network info cache due to event network-changed-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 871.106369] env[61995]: DEBUG oslo_concurrency.lockutils [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] Acquiring lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.106511] env[61995]: DEBUG oslo_concurrency.lockutils [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] Acquired lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.106705] env[61995]: DEBUG nova.network.neutron [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Refreshing network info cache for port cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.129644] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 871.129913] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 871.130159] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleting the datastore file [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.131394] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-199bf539-a76c-43bb-b5f6-68ecb5af0f4b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.137339] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52daad9c-f902-ec40-1d27-554fc2218c6c, 'name': SearchDatastore_Task, 'duration_secs': 0.017392} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.141991] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b73224f5-1840-4210-8ff5-869463e2fbb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.153286] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 871.153286] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]524abcae-1fd3-e886-fe2f-20a8caf63a33" [ 871.153286] env[61995]: _type = "Task" [ 871.153286] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.153801] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 871.153801] env[61995]: value = "task-794693" [ 871.153801] env[61995]: _type = "Task" [ 871.153801] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.172171] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524abcae-1fd3-e886-fe2f-20a8caf63a33, 'name': SearchDatastore_Task, 'duration_secs': 0.011919} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.175773] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.176168] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] ed952baf-f1ef-4ec0-8a15-928287190456/ed952baf-f1ef-4ec0-8a15-928287190456.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 871.176509] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794693, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.176778] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d17f375-9d5a-4c2b-b0b8-fa70b1328306 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.188288] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 871.188288] env[61995]: value = "task-794694" [ 871.188288] env[61995]: _type = "Task" [ 871.188288] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.206807] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794694, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.274170] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1819c81-73fd-4492-96ad-cb831c0b2ec6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.285158] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc738290-934f-4a21-ae2b-96475967e50f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.321446] env[61995]: DEBUG nova.network.neutron [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 871.323904] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a4f368-81f8-4679-8ff3-210a60ad5b1f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.333186] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0b98f6-5867-4faa-8fac-9f9745a5df19 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.349577] env[61995]: DEBUG nova.compute.provider_tree [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.477143] env[61995]: DEBUG nova.network.neutron [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.491195] env[61995]: DEBUG nova.network.neutron [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance_info_cache with network_info: [{"id": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "address": "fa:16:3e:45:26:f9", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb488f7-74", "ovs_interfaceid": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.609449] env[61995]: DEBUG nova.compute.manager [None req-1a44d52e-dc47-480c-8a53-743171c60d17 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Instance disappeared during snapshot {{(pid=61995) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 871.670660] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794693, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311741} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.672228] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.672373] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 871.672573] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 871.714046] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794694, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.855058] env[61995]: DEBUG nova.scheduler.client.report [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.868769] env[61995]: DEBUG nova.compute.manager [None req-1a44d52e-dc47-480c-8a53-743171c60d17 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Found 0 images (rotation: 2) {{(pid=61995) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 871.980575] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Releasing lock "refresh_cache-cfcf1a5a-97db-4095-b6a6-044242ee8a05" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.981036] env[61995]: DEBUG nova.compute.manager [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.981310] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 871.982144] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acea8a3e-0cb7-46a1-b0f4-d90ba4ea5abb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.993095] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 871.993095] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e3d1436-6999-41d1-afc8-af59c32e2d9d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.996406] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.997311] env[61995]: DEBUG nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Instance network_info: |[{"id": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "address": "fa:16:3e:45:26:f9", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb488f7-74", "ovs_interfaceid": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 871.997311] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:26:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9eb488f7-7474-4b91-9b01-2f4378f5daca', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.007312] env[61995]: DEBUG oslo.service.loopingcall [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.013028] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 872.013028] env[61995]: DEBUG oslo_vmware.api [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 872.013028] env[61995]: value = "task-794695" [ 872.013028] env[61995]: _type = "Task" [ 872.013028] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.013028] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1590471-9a86-416b-a4f6-b6f89400d674 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.046554] env[61995]: DEBUG oslo_vmware.api [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.048463] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.048463] env[61995]: value = "task-794696" [ 872.048463] env[61995]: _type = "Task" [ 872.048463] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.062602] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794696, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.202783] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794694, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.750629} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.203520] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] ed952baf-f1ef-4ec0-8a15-928287190456/ed952baf-f1ef-4ec0-8a15-928287190456.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 872.203520] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.204197] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2c0cdcb-b745-40fa-b050-4f5f7ef6e3be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.211019] env[61995]: DEBUG nova.network.neutron [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updated VIF entry in instance network info cache for port cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.211019] env[61995]: DEBUG nova.network.neutron [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.219300] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 872.219300] env[61995]: value = "task-794697" [ 872.219300] env[61995]: _type = "Task" [ 872.219300] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.234228] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794697, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.354988] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquiring lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.354988] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.364747] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.365371] env[61995]: DEBUG nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.368841] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 13.797s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.546905] env[61995]: DEBUG oslo_vmware.api [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794695, 'name': PowerOffVM_Task, 'duration_secs': 0.205658} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.547301] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 872.547522] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 872.547827] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2af3dc4-6d75-406f-8cf1-57cf2bdaf696 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.566361] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794696, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.591076] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 872.591076] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 872.591076] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Deleting the datastore file [datastore2] cfcf1a5a-97db-4095-b6a6-044242ee8a05 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.591076] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39dbd1c9-7e1f-42c3-a6fb-895a42b838a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.606429] env[61995]: DEBUG oslo_vmware.api [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for the task: (returnval){ [ 872.606429] env[61995]: value = "task-794699" [ 872.606429] env[61995]: _type = "Task" [ 872.606429] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.618479] env[61995]: DEBUG oslo_vmware.api [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794699, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.712905] env[61995]: DEBUG oslo_concurrency.lockutils [req-c1681143-f92d-4268-a544-a0b4d24d9cef req-dbd049ef-36ba-4408-bf04-e08ab464f4a1 service nova] Releasing lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.723537] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.723788] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.723951] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.724227] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.724337] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.724476] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.724696] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.724859] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.725040] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.725215] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.725410] env[61995]: DEBUG nova.virt.hardware [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.733056] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd09df4-9701-4ffb-bce7-203786fd842a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.741552] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794697, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078055} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.741884] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.743102] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ced7ab-e650-4812-98d3-65545f19ce60 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.747564] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbc22f0-20a7-448b-b9e5-6591e8f1598b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.793884] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] ed952baf-f1ef-4ec0-8a15-928287190456/ed952baf-f1ef-4ec0-8a15-928287190456.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.794524] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:68:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22b10d97-e8c2-4f19-b6b6-d5a55fe432c8', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.802221] env[61995]: DEBUG oslo.service.loopingcall [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.802321] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a399209-8394-495c-8b14-0f55c8dc72b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.816571] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 872.816829] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54e6dde6-8d07-4e36-82e8-0eed29441bcc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.839548] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 872.839548] env[61995]: value = "task-794700" [ 872.839548] env[61995]: _type = "Task" [ 872.839548] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.841589] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.841589] env[61995]: value = "task-794701" [ 872.841589] env[61995]: _type = "Task" [ 872.841589] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.854416] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794700, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.858290] env[61995]: DEBUG nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.861592] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794701, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.872838] env[61995]: DEBUG nova.compute.utils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.876110] env[61995]: INFO nova.compute.claims [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.880913] env[61995]: DEBUG nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.880913] env[61995]: DEBUG nova.network.neutron [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 873.027451] env[61995]: DEBUG nova.policy [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3efdbd0f8e0340dc8e4bfbf672d62f0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83064cbae17c429d8d084837635486da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 873.068195] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794696, 'name': CreateVM_Task, 'duration_secs': 0.605692} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.068573] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 873.069318] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.069516] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.070088] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.070189] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e50038e5-b7d3-4566-abfa-a9c131638613 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.082640] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 873.082640] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d91426-5fcf-6280-508f-8ca6a1c6eed8" [ 873.082640] env[61995]: _type = "Task" [ 873.082640] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.094941] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d91426-5fcf-6280-508f-8ca6a1c6eed8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.118228] env[61995]: DEBUG oslo_vmware.api [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Task: {'id': task-794699, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130319} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.118228] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.118325] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 873.118468] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 873.118695] env[61995]: INFO nova.compute.manager [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Took 1.14 seconds to destroy the instance on the hypervisor. [ 873.118913] env[61995]: DEBUG oslo.service.loopingcall [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.119203] env[61995]: DEBUG nova.compute.manager [-] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.119335] env[61995]: DEBUG nova.network.neutron [-] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 873.141778] env[61995]: DEBUG nova.network.neutron [-] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 873.356920] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794701, 'name': CreateVM_Task, 'duration_secs': 0.501} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.360919] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 873.360919] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794700, 'name': ReconfigVM_Task, 'duration_secs': 0.494709} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.360919] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.361323] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Reconfigured VM instance instance-00000042 to attach disk [datastore1] ed952baf-f1ef-4ec0-8a15-928287190456/ed952baf-f1ef-4ec0-8a15-928287190456.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.361680] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-584f1dd4-5e97-49c3-8dfa-1b6f8d813668 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.371021] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 873.371021] env[61995]: value = "task-794702" [ 873.371021] env[61995]: _type = "Task" [ 873.371021] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.386501] env[61995]: DEBUG nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.392109] env[61995]: INFO nova.compute.resource_tracker [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating resource usage from migration a8348ead-3c45-4010-98f6-841d2a6dec81 [ 873.397220] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794702, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.398920] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.593668] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d91426-5fcf-6280-508f-8ca6a1c6eed8, 'name': SearchDatastore_Task, 'duration_secs': 0.018869} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.593997] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.594245] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.594498] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.594647] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.594827] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.596440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.596440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.596440] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a50870b4-92cb-4a9f-8af6-53e19b308fa4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.600420] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f5474a3-8a40-4394-9a8d-a5d9bdf25fc8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.609844] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 873.609844] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e88070-a3f7-df33-25bc-29ae38d48ca2" [ 873.609844] env[61995]: _type = "Task" [ 873.609844] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.615971] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.615971] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 873.623488] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8513da9e-974a-49ce-9654-4523a60416c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.625837] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e88070-a3f7-df33-25bc-29ae38d48ca2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.627761] env[61995]: DEBUG nova.network.neutron [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Successfully created port: a410e471-f28f-4398-81c2-d4cd2bc69ba8 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.635225] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 873.635225] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523a5003-62de-7d5c-e549-e6d54f8c7c62" [ 873.635225] env[61995]: _type = "Task" [ 873.635225] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.644209] env[61995]: DEBUG nova.network.neutron [-] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.650151] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523a5003-62de-7d5c-e549-e6d54f8c7c62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.791364] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d046c44-4046-4613-99a7-7e21dc1b1b18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.801071] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0123406d-1d33-45f2-b164-9d98ad6b81f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.833307] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2c5dac-fab2-42b8-b3a1-75cfdce00d26 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.842028] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0309641-c209-42fe-9110-ea9c6096c354 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.856785] env[61995]: DEBUG nova.compute.provider_tree [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.882557] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794702, 'name': Rename_Task, 'duration_secs': 0.351673} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.882726] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 873.883440] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53c227dc-c287-4dd2-b74a-f2856b3db2ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.891681] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 873.891681] env[61995]: value = "task-794706" [ 873.891681] env[61995]: _type = "Task" [ 873.891681] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.911320] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.121791] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e88070-a3f7-df33-25bc-29ae38d48ca2, 'name': SearchDatastore_Task, 'duration_secs': 0.037043} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.121791] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.121944] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.122181] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.150339] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523a5003-62de-7d5c-e549-e6d54f8c7c62, 'name': SearchDatastore_Task, 'duration_secs': 0.019761} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.151366] env[61995]: INFO nova.compute.manager [-] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Took 1.03 seconds to deallocate network for instance. [ 874.151614] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db47bdf1-9707-4ff7-9052-8d8bea123d09 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.161433] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 874.161433] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52238581-f407-165b-6a4b-dd54445d7ff4" [ 874.161433] env[61995]: _type = "Task" [ 874.161433] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.169306] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52238581-f407-165b-6a4b-dd54445d7ff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.360829] env[61995]: DEBUG nova.scheduler.client.report [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.409176] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794706, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.410340] env[61995]: DEBUG nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.439847] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.440634] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.440879] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.441102] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.441269] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.441425] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.441641] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.441808] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.441980] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.442170] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.442349] env[61995]: DEBUG nova.virt.hardware [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.443557] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd99ae1-c6a4-470f-9ea9-b64beb24ea9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.456061] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92490c71-d4b5-4e4a-a308-c49e61ebe781 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.662056] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.672974] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52238581-f407-165b-6a4b-dd54445d7ff4, 'name': SearchDatastore_Task, 'duration_secs': 0.011219} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.673308] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.673513] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 874.673972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.674449] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.674449] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81c8b562-dd09-485a-9b49-d1c3871cf5a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.677045] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ebcd3dc-f611-4418-ac45-828b0b723004 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.685042] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 874.685042] env[61995]: value = "task-794707" [ 874.685042] env[61995]: _type = "Task" [ 874.685042] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.687949] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.687949] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 874.691581] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12104c48-877c-4b44-b443-b463f98eb622 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.700483] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794707, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.701864] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 874.701864] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f6df12-8123-2ed4-678d-8ab65d465f76" [ 874.701864] env[61995]: _type = "Task" [ 874.701864] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.714924] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f6df12-8123-2ed4-678d-8ab65d465f76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.868790] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.499s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.868790] env[61995]: INFO nova.compute.manager [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Migrating [ 874.875616] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.209s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.877915] env[61995]: INFO nova.compute.claims [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.895841] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.895841] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.910150] env[61995]: DEBUG oslo_vmware.api [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794706, 'name': PowerOnVM_Task, 'duration_secs': 0.928314} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.910450] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 874.910773] env[61995]: INFO nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Took 10.59 seconds to spawn the instance on the hypervisor. [ 874.910842] env[61995]: DEBUG nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.912308] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52952c8-e328-45d1-85e8-b903ba530c15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.204264] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794707, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505904} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.209253] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 875.209587] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.209879] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0efd54e-9711-45e2-8ab2-5785ec604b36 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.218124] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f6df12-8123-2ed4-678d-8ab65d465f76, 'name': SearchDatastore_Task, 'duration_secs': 0.020431} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.220231] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 875.220231] env[61995]: value = "task-794708" [ 875.220231] env[61995]: _type = "Task" [ 875.220231] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.220594] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a50a2fa-d9a1-4f4c-aa6c-db8764ef0b9e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.232807] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 875.232807] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529e7207-bb41-a4c1-6ad8-46160ef6e192" [ 875.232807] env[61995]: _type = "Task" [ 875.232807] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.236428] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.249633] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529e7207-bb41-a4c1-6ad8-46160ef6e192, 'name': SearchDatastore_Task, 'duration_secs': 0.011716} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.250031] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.250328] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 875.250594] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d6a0793-d5e0-4d0c-9545-bbe1ca285c69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.259544] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 875.259544] env[61995]: value = "task-794709" [ 875.259544] env[61995]: _type = "Task" [ 875.259544] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.270821] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.295867] env[61995]: DEBUG nova.compute.manager [req-b45c71f8-d7ed-41b3-9aff-d0fb5432ec51 req-bde01957-2ece-4a40-9ea7-ee2bfec628c5 service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Received event network-vif-plugged-a410e471-f28f-4398-81c2-d4cd2bc69ba8 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.296140] env[61995]: DEBUG oslo_concurrency.lockutils [req-b45c71f8-d7ed-41b3-9aff-d0fb5432ec51 req-bde01957-2ece-4a40-9ea7-ee2bfec628c5 service nova] Acquiring lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.296390] env[61995]: DEBUG oslo_concurrency.lockutils [req-b45c71f8-d7ed-41b3-9aff-d0fb5432ec51 req-bde01957-2ece-4a40-9ea7-ee2bfec628c5 service nova] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.296567] env[61995]: DEBUG oslo_concurrency.lockutils [req-b45c71f8-d7ed-41b3-9aff-d0fb5432ec51 req-bde01957-2ece-4a40-9ea7-ee2bfec628c5 service nova] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.296742] env[61995]: DEBUG nova.compute.manager [req-b45c71f8-d7ed-41b3-9aff-d0fb5432ec51 req-bde01957-2ece-4a40-9ea7-ee2bfec628c5 service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] No waiting events found dispatching network-vif-plugged-a410e471-f28f-4398-81c2-d4cd2bc69ba8 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.296979] env[61995]: WARNING nova.compute.manager [req-b45c71f8-d7ed-41b3-9aff-d0fb5432ec51 req-bde01957-2ece-4a40-9ea7-ee2bfec628c5 service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Received unexpected event network-vif-plugged-a410e471-f28f-4398-81c2-d4cd2bc69ba8 for instance with vm_state building and task_state spawning. [ 875.398465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.398465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.398465] env[61995]: DEBUG nova.network.neutron [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.398465] env[61995]: DEBUG nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 875.437447] env[61995]: DEBUG nova.network.neutron [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Successfully updated port: a410e471-f28f-4398-81c2-d4cd2bc69ba8 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.444103] env[61995]: INFO nova.compute.manager [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Took 34.05 seconds to build instance. [ 875.736928] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070308} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.737252] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.738108] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da83569-bd30-428f-a980-8ec9205ccbff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.762878] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.762878] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-942b9d4b-0008-452d-835a-9aed53def7f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.788186] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47873} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.789768] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 875.790279] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.790460] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 875.790460] env[61995]: value = "task-794710" [ 875.790460] env[61995]: _type = "Task" [ 875.790460] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.790685] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-683e879a-c33e-4ee8-87c9-5fd0b6fe0e59 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.802938] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.804460] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 875.804460] env[61995]: value = "task-794711" [ 875.804460] env[61995]: _type = "Task" [ 875.804460] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.817555] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794711, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.929947] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.943360] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "refresh_cache-fdeceddf-c6f9-4a98-afd0-defa1284eeff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.943511] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "refresh_cache-fdeceddf-c6f9-4a98-afd0-defa1284eeff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.943685] env[61995]: DEBUG nova.network.neutron [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.950021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e06a6024-b8c3-4a73-a8ec-02dc85d16b00 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "ed952baf-f1ef-4ec0-8a15-928287190456" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.560s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.255353] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9850bad-ff89-4f4b-9502-c4eec3e3adf3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.264184] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bafee4-d56a-4a29-9914-033b80ba4ad3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.306223] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544e5c86-3ccf-4698-9f25-ba5cd32648af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.317847] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794711, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07726} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.324147] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.324681] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794710, 'name': ReconfigVM_Task, 'duration_secs': 0.343245} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.325754] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d98056-5ad7-4437-adc3-6d820d88d7de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.328454] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.330198] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faed12eb-33ef-4380-a082-4f2476919387 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.334889] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2458bd78-00d7-468c-8c02-f26667436032 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.350401] env[61995]: DEBUG nova.compute.provider_tree [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.364324] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 876.364324] env[61995]: value = "task-794712" [ 876.364324] env[61995]: _type = "Task" [ 876.364324] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.373354] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.374754] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d462c84f-88c4-4a69-a73a-f0f04bf4288d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.407794] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794712, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.410349] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 876.410349] env[61995]: value = "task-794713" [ 876.410349] env[61995]: _type = "Task" [ 876.410349] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.414646] env[61995]: DEBUG nova.network.neutron [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance_info_cache with network_info: [{"id": "3ecd14b2-d6b9-4216-936e-edae029037af", "address": "fa:16:3e:2b:38:33", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ecd14b2-d6", "ovs_interfaceid": "3ecd14b2-d6b9-4216-936e-edae029037af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.430707] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794713, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.500111] env[61995]: DEBUG nova.network.neutron [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 876.739344] env[61995]: DEBUG nova.network.neutron [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Updating instance_info_cache with network_info: [{"id": "a410e471-f28f-4398-81c2-d4cd2bc69ba8", "address": "fa:16:3e:a9:90:d1", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa410e471-f2", "ovs_interfaceid": "a410e471-f28f-4398-81c2-d4cd2bc69ba8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.875451] env[61995]: DEBUG nova.scheduler.client.report [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.896361] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794712, 'name': Rename_Task, 'duration_secs': 0.189652} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.896697] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 876.896987] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c554da2-d350-4a72-9de3-e427d15310c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.907603] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 876.907603] env[61995]: value = "task-794715" [ 876.907603] env[61995]: _type = "Task" [ 876.907603] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.927631] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.929472] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794715, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.929746] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794713, 'name': ReconfigVM_Task, 'duration_secs': 0.503358} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.930666] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Reconfigured VM instance instance-00000024 to attach disk [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.931483] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dac9ce5e-1a0f-4bec-9680-d1097b30e2a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.943066] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 876.943066] env[61995]: value = "task-794716" [ 876.943066] env[61995]: _type = "Task" [ 876.943066] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.959184] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794716, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.243295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "refresh_cache-fdeceddf-c6f9-4a98-afd0-defa1284eeff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.243439] env[61995]: DEBUG nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Instance network_info: |[{"id": "a410e471-f28f-4398-81c2-d4cd2bc69ba8", "address": "fa:16:3e:a9:90:d1", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa410e471-f2", "ovs_interfaceid": "a410e471-f28f-4398-81c2-d4cd2bc69ba8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.243899] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:90:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a410e471-f28f-4398-81c2-d4cd2bc69ba8', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.253106] env[61995]: DEBUG oslo.service.loopingcall [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.253423] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 877.253717] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6cac675e-4e35-443a-918f-1cc86217c411 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.274241] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.274241] env[61995]: value = "task-794717" [ 877.274241] env[61995]: _type = "Task" [ 877.274241] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.284269] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794717, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.391017] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.392031] env[61995]: DEBUG nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.395799] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.671s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.396073] env[61995]: DEBUG nova.objects.instance [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lazy-loading 'resources' on Instance uuid bb31ef64-478f-4f39-9ee2-fcfabe8235ec {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.402168] env[61995]: DEBUG nova.compute.manager [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Received event network-changed-a410e471-f28f-4398-81c2-d4cd2bc69ba8 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.402168] env[61995]: DEBUG nova.compute.manager [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Refreshing instance network info cache due to event network-changed-a410e471-f28f-4398-81c2-d4cd2bc69ba8. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 877.402168] env[61995]: DEBUG oslo_concurrency.lockutils [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] Acquiring lock "refresh_cache-fdeceddf-c6f9-4a98-afd0-defa1284eeff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.402168] env[61995]: DEBUG oslo_concurrency.lockutils [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] Acquired lock "refresh_cache-fdeceddf-c6f9-4a98-afd0-defa1284eeff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.402168] env[61995]: DEBUG nova.network.neutron [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Refreshing network info cache for port a410e471-f28f-4398-81c2-d4cd2bc69ba8 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 877.422096] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794715, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.461281] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794716, 'name': Rename_Task, 'duration_secs': 0.319202} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.461281] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 877.461281] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74712f56-e794-4bc9-bcc0-fb435db3c2cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.469367] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 877.469367] env[61995]: value = "task-794718" [ 877.469367] env[61995]: _type = "Task" [ 877.469367] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.479866] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794718, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.785694] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794717, 'name': CreateVM_Task, 'duration_secs': 0.501672} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.785916] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 877.786726] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.786919] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.787615] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.787919] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d97e6bf-a28b-4093-8163-4f53a219b2ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.793567] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 877.793567] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525337c9-8d63-6ab8-e512-835b9aab8681" [ 877.793567] env[61995]: _type = "Task" [ 877.793567] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.803409] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525337c9-8d63-6ab8-e512-835b9aab8681, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.908802] env[61995]: DEBUG nova.compute.utils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.914038] env[61995]: DEBUG nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.914038] env[61995]: DEBUG nova.network.neutron [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 877.929727] env[61995]: DEBUG oslo_vmware.api [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794715, 'name': PowerOnVM_Task, 'duration_secs': 0.563034} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.930011] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 877.930239] env[61995]: INFO nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Took 11.04 seconds to spawn the instance on the hypervisor. [ 877.930462] env[61995]: DEBUG nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.931562] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1076beef-a824-4361-9b84-b0cc9ee34de8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.986445] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794718, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.029951] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.030052] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.060465] env[61995]: DEBUG nova.policy [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f27aa509c254e2fa6a5d833b53f50e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4042c97b25a24bf0a1ebe23e958f95ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 878.289063] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed371b85-cdd4-4c7b-b184-73447ca66a12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.299678] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519d54c2-e47c-47a1-8f97-8b61c406960f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.337740] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525337c9-8d63-6ab8-e512-835b9aab8681, 'name': SearchDatastore_Task, 'duration_secs': 0.011467} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.341325] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.341776] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.342178] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.342496] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.342823] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.343711] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5f0be2-823d-4215-8a19-3c0ebcf7d7fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.349048] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f790ee75-6632-4cdb-8d32-f8d64ba49c9f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.355095] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426623a5-31a0-4ab1-88d6-4786bb5543e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.361258] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.361667] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 878.362814] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3404c0fc-68b8-4993-be4c-4e069e139285 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.373270] env[61995]: DEBUG nova.compute.provider_tree [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.378206] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 878.378206] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529c20e4-d2d1-183f-5404-04c8b66b72bd" [ 878.378206] env[61995]: _type = "Task" [ 878.378206] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.387263] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529c20e4-d2d1-183f-5404-04c8b66b72bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.416419] env[61995]: DEBUG nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.445726] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f894207f-84a3-4442-b33c-e62ff69f7d37 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.474918] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance '2b40609a-0826-4a16-af28-c8926af6b646' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 878.478704] env[61995]: INFO nova.compute.manager [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Took 34.29 seconds to build instance. [ 878.489932] env[61995]: DEBUG oslo_vmware.api [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794718, 'name': PowerOnVM_Task, 'duration_secs': 0.545891} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.490355] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.490444] env[61995]: DEBUG nova.compute.manager [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.491514] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f3e895-6253-49c3-898e-77f01189fce7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.537825] env[61995]: DEBUG nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 878.578328] env[61995]: DEBUG nova.network.neutron [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Updated VIF entry in instance network info cache for port a410e471-f28f-4398-81c2-d4cd2bc69ba8. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 878.578707] env[61995]: DEBUG nova.network.neutron [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Updating instance_info_cache with network_info: [{"id": "a410e471-f28f-4398-81c2-d4cd2bc69ba8", "address": "fa:16:3e:a9:90:d1", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa410e471-f2", "ovs_interfaceid": "a410e471-f28f-4398-81c2-d4cd2bc69ba8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.878584] env[61995]: DEBUG nova.scheduler.client.report [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.891134] env[61995]: DEBUG nova.network.neutron [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Successfully created port: af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.902931] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529c20e4-d2d1-183f-5404-04c8b66b72bd, 'name': SearchDatastore_Task, 'duration_secs': 0.023101} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.905843] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47927254-7f1e-4957-9c07-5ff5881ff30c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.914987] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 878.914987] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ed7644-ec29-d822-97f6-03152ec21928" [ 878.914987] env[61995]: _type = "Task" [ 878.914987] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.931157] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ed7644-ec29-d822-97f6-03152ec21928, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.985248] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 878.986233] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ecc7c4a5-9999-4f45-a590-5da045cf0cb6 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.804s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.986823] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71e5341d-6b96-4844-a218-b891191a29d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.995873] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 878.995873] env[61995]: value = "task-794720" [ 878.995873] env[61995]: _type = "Task" [ 878.995873] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.015723] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.019999] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.071405] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.082535] env[61995]: DEBUG oslo_concurrency.lockutils [req-e8de4b90-82ad-41a9-99d0-72273b6162f9 req-2cd05a2e-bd47-4034-a41b-cb424ffa12bd service nova] Releasing lock "refresh_cache-fdeceddf-c6f9-4a98-afd0-defa1284eeff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.385416] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.388691] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.223s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.388691] env[61995]: DEBUG nova.objects.instance [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lazy-loading 'resources' on Instance uuid af7163e8-4e77-4336-80e9-0623e29ed8c0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.410349] env[61995]: INFO nova.scheduler.client.report [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleted allocations for instance bb31ef64-478f-4f39-9ee2-fcfabe8235ec [ 879.431435] env[61995]: DEBUG nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.433964] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ed7644-ec29-d822-97f6-03152ec21928, 'name': SearchDatastore_Task, 'duration_secs': 0.018704} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.434853] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.435136] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 879.435488] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76a307de-1de8-4fbf-befc-d4d1cfb30787 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.444285] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 879.444285] env[61995]: value = "task-794721" [ 879.444285] env[61995]: _type = "Task" [ 879.444285] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.453236] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794721, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.472797] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.474025] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.474025] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.474274] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.474446] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.474725] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.475071] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.475362] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.475638] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.475836] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.476317] env[61995]: DEBUG nova.virt.hardware [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.477737] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6a0835-9464-46de-8f13-9ac8cab1c62f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.492349] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5d5934-cff8-4de2-8495-73933bbf058d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.516896] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794720, 'name': PowerOffVM_Task, 'duration_secs': 0.373027} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.517658] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 879.517658] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance '2b40609a-0826-4a16-af28-c8926af6b646' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 879.919553] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2af1072d-db12-4480-983f-b7d00f9e6a23 tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "bb31ef64-478f-4f39-9ee2-fcfabe8235ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.505s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.954703] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794721, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48103} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.955070] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 879.955298] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.955587] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b9fce9e-cdbf-4fbf-bef0-d36390f274ca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.971638] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 879.971638] env[61995]: value = "task-794722" [ 879.971638] env[61995]: _type = "Task" [ 879.971638] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.987678] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794722, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.029260] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.029587] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.029773] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.029963] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.030145] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.030302] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.030512] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.030694] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.030873] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.031058] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.031270] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.040590] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ed797aa-522c-4c16-869c-cf50984d9c95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.059367] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 880.059367] env[61995]: value = "task-794723" [ 880.059367] env[61995]: _type = "Task" [ 880.059367] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.069352] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794723, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.286790] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eaf830d-dd15-4a5c-a22f-32cf73221b61 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.296413] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59168b07-c051-475b-958a-bb9982867d96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.330909] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7685b90-dc8c-4bfb-b078-9708d4ac54f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.335397] env[61995]: INFO nova.compute.manager [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Rebuilding instance [ 880.344515] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2231fbfe-d852-4de9-bf7f-a695630cd3fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.363958] env[61995]: DEBUG nova.compute.provider_tree [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.395019] env[61995]: DEBUG nova.compute.manager [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.395019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead21dce-02c5-4e3f-85a2-a6f2002ec0c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.484458] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794722, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090795} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.485304] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.486247] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb28d639-bce8-4ce6-9112-177c93ac0938 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.514027] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.514027] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-483e7df5-0787-42b6-afab-74fcdcb8b457 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.536279] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 880.536279] env[61995]: value = "task-794724" [ 880.536279] env[61995]: _type = "Task" [ 880.536279] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.547660] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794724, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.570358] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794723, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.868907] env[61995]: DEBUG nova.scheduler.client.report [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.907734] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 880.907943] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2767ff5f-c553-45c5-a6ce-0d7c21363aa4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.917564] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 880.917564] env[61995]: value = "task-794726" [ 880.917564] env[61995]: _type = "Task" [ 880.917564] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.931665] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794726, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.984020] env[61995]: DEBUG nova.compute.manager [req-0d798abe-aabd-4174-9643-3ac44657cdf9 req-e47aaf2c-52a5-4e66-a086-e2658fdc785d service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Received event network-vif-plugged-af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.984397] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d798abe-aabd-4174-9643-3ac44657cdf9 req-e47aaf2c-52a5-4e66-a086-e2658fdc785d service nova] Acquiring lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.984615] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d798abe-aabd-4174-9643-3ac44657cdf9 req-e47aaf2c-52a5-4e66-a086-e2658fdc785d service nova] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.984808] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d798abe-aabd-4174-9643-3ac44657cdf9 req-e47aaf2c-52a5-4e66-a086-e2658fdc785d service nova] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.985044] env[61995]: DEBUG nova.compute.manager [req-0d798abe-aabd-4174-9643-3ac44657cdf9 req-e47aaf2c-52a5-4e66-a086-e2658fdc785d service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] No waiting events found dispatching network-vif-plugged-af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.985416] env[61995]: WARNING nova.compute.manager [req-0d798abe-aabd-4174-9643-3ac44657cdf9 req-e47aaf2c-52a5-4e66-a086-e2658fdc785d service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Received unexpected event network-vif-plugged-af010e12-84c9-4a5d-b901-4f519c51ce79 for instance with vm_state building and task_state spawning. [ 881.048194] env[61995]: DEBUG nova.network.neutron [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Successfully updated port: af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.056849] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794724, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.073578] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794723, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.383580] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.995s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.389009] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.990s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.392199] env[61995]: INFO nova.compute.claims [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.403018] env[61995]: DEBUG nova.compute.manager [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 881.427169] env[61995]: INFO nova.scheduler.client.report [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Deleted allocations for instance af7163e8-4e77-4336-80e9-0623e29ed8c0 [ 881.439300] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794726, 'name': PowerOffVM_Task, 'duration_secs': 0.289656} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.439471] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 881.439717] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 881.440641] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d83b6d-1224-44d4-a743-12eb023c5c08 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.450843] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 881.451193] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af5c3d76-6e38-42c0-b11b-fbf48c8f6a69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.548391] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794724, 'name': ReconfigVM_Task, 'duration_secs': 0.617691} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.548694] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Reconfigured VM instance instance-00000044 to attach disk [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.549490] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32117653-86f5-4594-863d-e6c41917d700 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.557122] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 881.557122] env[61995]: value = "task-794728" [ 881.557122] env[61995]: _type = "Task" [ 881.557122] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.557699] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.557843] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.558436] env[61995]: DEBUG nova.network.neutron [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 881.571798] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794728, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.577098] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794723, 'name': ReconfigVM_Task, 'duration_secs': 1.298762} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.577441] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance '2b40609a-0826-4a16-af28-c8926af6b646' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 881.816296] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 881.816296] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 881.816296] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleting the datastore file [datastore1] c4f744f9-36f3-4913-9ced-f0db93d3b4df {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.816296] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8cf2a67c-a86b-4a96-a950-d2893486a91a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.823063] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 881.823063] env[61995]: value = "task-794729" [ 881.823063] env[61995]: _type = "Task" [ 881.823063] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.836048] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.927288] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.937685] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a45876b-c072-4a60-a62a-0876bb60400b tempest-ListServersNegativeTestJSON-74386100 tempest-ListServersNegativeTestJSON-74386100-project-member] Lock "af7163e8-4e77-4336-80e9-0623e29ed8c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.460s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.069549] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794728, 'name': Rename_Task, 'duration_secs': 0.174879} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.069549] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 882.069549] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5dde9451-4d21-4d2b-9ff1-88f5f37e183d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.076777] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 882.076777] env[61995]: value = "task-794730" [ 882.076777] env[61995]: _type = "Task" [ 882.076777] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.087533] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.087930] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.088154] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.088350] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.088498] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.088675] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.089042] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.089324] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.089513] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.089681] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.090582] env[61995]: DEBUG nova.virt.hardware [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.095411] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Reconfiguring VM instance instance-0000003e to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 882.095625] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794730, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.095917] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06fc6aa4-4a45-4899-9d88-6d0807fa4a84 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.109989] env[61995]: DEBUG nova.network.neutron [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.120523] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 882.120523] env[61995]: value = "task-794731" [ 882.120523] env[61995]: _type = "Task" [ 882.120523] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.130774] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794731, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.343221] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157634} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.343221] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.343221] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 882.343221] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 882.476272] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.476536] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.587632] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794730, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.635247] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794731, 'name': ReconfigVM_Task, 'duration_secs': 0.272696} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.635247] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Reconfigured VM instance instance-0000003e to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 882.635899] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7987b636-ce86-444f-adc8-cb3efaf23a49 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.660252] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 2b40609a-0826-4a16-af28-c8926af6b646/2b40609a-0826-4a16-af28-c8926af6b646.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.663187] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f6b2c11-4bf1-4b52-9972-16c5d3f49989 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.678947] env[61995]: DEBUG nova.network.neutron [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Updating instance_info_cache with network_info: [{"id": "af010e12-84c9-4a5d-b901-4f519c51ce79", "address": "fa:16:3e:c8:6c:9a", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf010e12-84", "ovs_interfaceid": "af010e12-84c9-4a5d-b901-4f519c51ce79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.688533] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 882.688533] env[61995]: value = "task-794732" [ 882.688533] env[61995]: _type = "Task" [ 882.688533] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.702649] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794732, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.802221] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44253776-7929-4ac0-8e1f-980e41c90e53 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.812468] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf609bae-f1f3-487f-8b51-881d78f495f3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.853115] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51622295-0964-4c16-bc9f-ae6e837dcbd6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.861473] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead5927b-f2e9-401a-b121-ccb2e0d69dee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.878830] env[61995]: DEBUG nova.compute.provider_tree [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.982027] env[61995]: DEBUG nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.095022] env[61995]: DEBUG oslo_vmware.api [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794730, 'name': PowerOnVM_Task, 'duration_secs': 0.751095} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.095022] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 883.095022] env[61995]: INFO nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Took 8.68 seconds to spawn the instance on the hypervisor. [ 883.095022] env[61995]: DEBUG nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.095022] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a37d41-0105-4cb8-bbce-697a51c051a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.184815] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.184815] env[61995]: DEBUG nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Instance network_info: |[{"id": "af010e12-84c9-4a5d-b901-4f519c51ce79", "address": "fa:16:3e:c8:6c:9a", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf010e12-84", "ovs_interfaceid": "af010e12-84c9-4a5d-b901-4f519c51ce79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.184815] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:6c:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3291573-fad8-48cc-a965-c3554e7cee4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af010e12-84c9-4a5d-b901-4f519c51ce79', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.194187] env[61995]: DEBUG oslo.service.loopingcall [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.194930] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 883.198856] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5cbfedbc-6650-4ddc-bb9a-637e5fbc1a55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.220737] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794732, 'name': ReconfigVM_Task, 'duration_secs': 0.505637} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.222423] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 2b40609a-0826-4a16-af28-c8926af6b646/2b40609a-0826-4a16-af28-c8926af6b646.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.222805] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance '2b40609a-0826-4a16-af28-c8926af6b646' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 883.226186] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.226186] env[61995]: value = "task-794734" [ 883.226186] env[61995]: _type = "Task" [ 883.226186] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.236848] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794734, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.239626] env[61995]: DEBUG nova.compute.manager [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Received event network-changed-af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.239626] env[61995]: DEBUG nova.compute.manager [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Refreshing instance network info cache due to event network-changed-af010e12-84c9-4a5d-b901-4f519c51ce79. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.239800] env[61995]: DEBUG oslo_concurrency.lockutils [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] Acquiring lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.239955] env[61995]: DEBUG oslo_concurrency.lockutils [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] Acquired lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.240142] env[61995]: DEBUG nova.network.neutron [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Refreshing network info cache for port af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.276617] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.276990] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.277189] env[61995]: DEBUG nova.compute.manager [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.279131] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db2f5fa-7033-4bfe-abe5-b0d502f826d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.289638] env[61995]: DEBUG nova.compute.manager [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 883.290379] env[61995]: DEBUG nova.objects.instance [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'flavor' on Instance uuid d9c53806-68c0-4872-a262-36bc05573674 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.382347] env[61995]: DEBUG nova.scheduler.client.report [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.388373] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.388801] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.389059] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.389287] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.389485] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.389844] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.390053] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.390292] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.390612] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.391012] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.391391] env[61995]: DEBUG nova.virt.hardware [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.392324] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53a97a9-63b3-40f2-a6f7-cd113ba84e6c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.403348] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328df5cc-cad2-42da-aa5b-1d850cde9b3c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.420734] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:68:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22b10d97-e8c2-4f19-b6b6-d5a55fe432c8', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.428565] env[61995]: DEBUG oslo.service.loopingcall [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.429269] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 883.429460] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8114b7bf-aeaa-4012-94b6-e43577a049b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.451627] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.451627] env[61995]: value = "task-794735" [ 883.451627] env[61995]: _type = "Task" [ 883.451627] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.460530] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794735, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.511760] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.618647] env[61995]: INFO nova.compute.manager [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Took 27.71 seconds to build instance. [ 883.732858] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bffc30b-181b-4cf4-a6c3-6813f457e2a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.742442] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794734, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.759807] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b53c3d-077a-4bb5-a4d4-d8524af7a5ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.784960] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance '2b40609a-0826-4a16-af28-c8926af6b646' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 883.796300] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.796665] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84926eb2-b869-4318-b141-b873c6b513bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.808024] env[61995]: DEBUG oslo_vmware.api [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 883.808024] env[61995]: value = "task-794736" [ 883.808024] env[61995]: _type = "Task" [ 883.808024] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.821908] env[61995]: DEBUG oslo_vmware.api [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794736, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.902479] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.902479] env[61995]: DEBUG nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.904157] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.242s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.904762] env[61995]: DEBUG nova.objects.instance [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lazy-loading 'resources' on Instance uuid cfcf1a5a-97db-4095-b6a6-044242ee8a05 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.961981] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794735, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.124185] env[61995]: DEBUG oslo_concurrency.lockutils [None req-878c9683-cd05-48cb-b9b8-4ffaed8cd613 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.221s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.237688] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794734, 'name': CreateVM_Task, 'duration_secs': 0.526325} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.237852] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 884.238908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.239147] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.239474] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.239730] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a3d9a97-d27b-4a75-a096-570135386fe9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.245174] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 884.245174] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5220b6ae-e911-2644-b572-6eecc8c8ebd7" [ 884.245174] env[61995]: _type = "Task" [ 884.245174] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.254145] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5220b6ae-e911-2644-b572-6eecc8c8ebd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.296861] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.297586] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.297586] env[61995]: DEBUG nova.compute.manager [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.298713] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87a90c9-79c1-463b-b451-73e5724af811 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.309029] env[61995]: DEBUG nova.compute.manager [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 884.310247] env[61995]: DEBUG nova.objects.instance [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'flavor' on Instance uuid fdeceddf-c6f9-4a98-afd0-defa1284eeff {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.314833] env[61995]: DEBUG nova.network.neutron [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Updated VIF entry in instance network info cache for port af010e12-84c9-4a5d-b901-4f519c51ce79. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.315222] env[61995]: DEBUG nova.network.neutron [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Updating instance_info_cache with network_info: [{"id": "af010e12-84c9-4a5d-b901-4f519c51ce79", "address": "fa:16:3e:c8:6c:9a", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf010e12-84", "ovs_interfaceid": "af010e12-84c9-4a5d-b901-4f519c51ce79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.329241] env[61995]: DEBUG oslo_vmware.api [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794736, 'name': PowerOffVM_Task, 'duration_secs': 0.292276} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.329960] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 884.330611] env[61995]: DEBUG nova.compute.manager [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.332258] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be9b1cb-dc13-4483-83c0-41a33f60dd0b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.376035] env[61995]: DEBUG nova.network.neutron [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Port 3ecd14b2-d6b9-4216-936e-edae029037af binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 884.409056] env[61995]: DEBUG nova.compute.utils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.413628] env[61995]: DEBUG nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.414471] env[61995]: DEBUG nova.network.neutron [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 884.469451] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794735, 'name': CreateVM_Task, 'duration_secs': 0.594964} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.472248] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 884.472941] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.477287] env[61995]: DEBUG nova.policy [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '710433964ecd44a58adb5dd50b47d950', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df7be4efd5314088a9df13b042f7b100', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 884.762212] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5220b6ae-e911-2644-b572-6eecc8c8ebd7, 'name': SearchDatastore_Task, 'duration_secs': 0.028934} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.762567] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.762783] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.767368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.767368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.767368] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.767368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.767368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.767368] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcda221b-a272-4536-b244-75eb368b777a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.767368] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-621e48f4-46e3-4af0-878a-84f0ba9e17a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.779947] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 884.779947] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52150e65-0700-108c-842a-a05899954701" [ 884.779947] env[61995]: _type = "Task" [ 884.779947] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.784485] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.784851] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 884.786472] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a54a7e-606b-4fd3-b67b-468b8c757821 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.789331] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f258b74-a223-4e16-b12e-3780d20edc7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.798179] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52150e65-0700-108c-842a-a05899954701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.804936] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 884.804936] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529de79e-7662-0d41-18e5-eccbc36af59b" [ 884.804936] env[61995]: _type = "Task" [ 884.804936] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.806042] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d96df67-817a-4f94-8ef9-c23dba338818 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.818453] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529de79e-7662-0d41-18e5-eccbc36af59b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.849708] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 884.850182] env[61995]: DEBUG oslo_concurrency.lockutils [req-c8ece270-8955-4f78-8867-15c51d93e7e0 req-794a3e12-a439-453e-ab1e-518b7e56c3fc service nova] Releasing lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.852630] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14c61e4c-5a98-4f81-96f9-98d46b356e63 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.854821] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eba970-46cb-4390-9cc8-2d7befc31de4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.860284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1723f03d-ef2f-4b9e-92bc-8ab7e29c9f74 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.583s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.868667] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e46644-033b-47d4-9105-a0fac645c638 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.874790] env[61995]: DEBUG oslo_vmware.api [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 884.874790] env[61995]: value = "task-794737" [ 884.874790] env[61995]: _type = "Task" [ 884.874790] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.892250] env[61995]: DEBUG nova.compute.provider_tree [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.901648] env[61995]: DEBUG oslo_vmware.api [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.914212] env[61995]: DEBUG nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.919213] env[61995]: DEBUG nova.network.neutron [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Successfully created port: ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.292210] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52150e65-0700-108c-842a-a05899954701, 'name': SearchDatastore_Task, 'duration_secs': 0.027357} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.292600] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.292712] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.292929] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.324583] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529de79e-7662-0d41-18e5-eccbc36af59b, 'name': SearchDatastore_Task, 'duration_secs': 0.017517} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.325259] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5307255-c911-4098-bc6b-7fa3a5468069 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.332672] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 885.332672] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ddc419-69b7-ab3d-3bb3-52cba5295134" [ 885.332672] env[61995]: _type = "Task" [ 885.332672] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.342554] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ddc419-69b7-ab3d-3bb3-52cba5295134, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.388342] env[61995]: DEBUG oslo_vmware.api [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794737, 'name': PowerOffVM_Task, 'duration_secs': 0.233168} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.388342] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 885.388342] env[61995]: DEBUG nova.compute.manager [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.388949] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d03453f-1c41-489a-889a-a16c36505640 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.404525] env[61995]: DEBUG nova.scheduler.client.report [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.417904] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "2b40609a-0826-4a16-af28-c8926af6b646-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.418373] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.419068] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.426712] env[61995]: INFO nova.virt.block_device [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Booting with volume 604ce120-96fa-4557-9558-055ca5966123 at /dev/sda [ 885.475622] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11e56273-ad46-4992-860e-36ef8bf6b01b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.486894] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8110d266-45f0-4d2c-b070-7fdd5a9905b6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.541615] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5bdb76ad-16a0-45fe-a9f7-cf75df6f4c1a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.551369] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd401f19-8bf0-413b-a9f3-e1d674fc9a47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.588587] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb361821-1b4b-4490-af46-5a8b38a90ed9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.596355] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf92a6ca-9b50-4eb3-a0ff-f6dbc4f87700 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.613186] env[61995]: DEBUG nova.virt.block_device [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updating existing volume attachment record: e3d2a1cb-08e2-472a-89ab-1e13e2668a8c {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 885.847638] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ddc419-69b7-ab3d-3bb3-52cba5295134, 'name': SearchDatastore_Task, 'duration_secs': 0.012279} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.847912] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.847960] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5/a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 885.848329] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.848592] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.848817] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ecb1a2bd-3831-4e0d-81f1-d1889ae27937 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.851106] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e404f5e8-3c01-4242-b524-1a6db842166a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.859206] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 885.859206] env[61995]: value = "task-794739" [ 885.859206] env[61995]: _type = "Task" [ 885.859206] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.863982] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.864119] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 885.865218] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92b4b0b2-16cb-4b5b-aea2-b9a0a0212ade {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.870891] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794739, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.875365] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 885.875365] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5261881e-f328-d6e5-1d38-d16975672c08" [ 885.875365] env[61995]: _type = "Task" [ 885.875365] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.884789] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5261881e-f328-d6e5-1d38-d16975672c08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.918941] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.015s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.928081] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b967a1ec-f000-437e-8c43-8b98f49de4b7 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.628s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.928081] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.997s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.928506] env[61995]: INFO nova.compute.claims [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.951112] env[61995]: INFO nova.scheduler.client.report [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Deleted allocations for instance cfcf1a5a-97db-4095-b6a6-044242ee8a05 [ 886.116519] env[61995]: DEBUG nova.objects.instance [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'flavor' on Instance uuid d9c53806-68c0-4872-a262-36bc05573674 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.370097] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794739, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465002} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.370097] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5/a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 886.370097] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.370378] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4bd6de52-fc44-4a1c-9c05-fad0d592a65e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.382984] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 886.382984] env[61995]: value = "task-794740" [ 886.382984] env[61995]: _type = "Task" [ 886.382984] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.391292] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5261881e-f328-d6e5-1d38-d16975672c08, 'name': SearchDatastore_Task, 'duration_secs': 0.009858} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.392569] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75a4cda8-ae6b-4a83-b48f-319b63ee58dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.399434] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.403603] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 886.403603] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52fdc21b-cd20-fa16-a7cf-93665cd9ed71" [ 886.403603] env[61995]: _type = "Task" [ 886.403603] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.411912] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fdc21b-cd20-fa16-a7cf-93665cd9ed71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.460065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a2b71722-889e-4778-96cd-0c8d947fe4e7 tempest-ServersAaction247Test-1375916088 tempest-ServersAaction247Test-1375916088-project-member] Lock "cfcf1a5a-97db-4095-b6a6-044242ee8a05" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.696s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.476418] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.476418] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.476418] env[61995]: DEBUG nova.network.neutron [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.622539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.622731] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.622951] env[61995]: DEBUG nova.network.neutron [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.623170] env[61995]: DEBUG nova.objects.instance [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'info_cache' on Instance uuid d9c53806-68c0-4872-a262-36bc05573674 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.821445] env[61995]: DEBUG nova.compute.manager [req-52e300ba-eda7-4192-8654-76f627c0879e req-3538b945-7e00-4ed3-b1f3-b593451059d4 service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Received event network-vif-plugged-ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.821727] env[61995]: DEBUG oslo_concurrency.lockutils [req-52e300ba-eda7-4192-8654-76f627c0879e req-3538b945-7e00-4ed3-b1f3-b593451059d4 service nova] Acquiring lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.822009] env[61995]: DEBUG oslo_concurrency.lockutils [req-52e300ba-eda7-4192-8654-76f627c0879e req-3538b945-7e00-4ed3-b1f3-b593451059d4 service nova] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.822389] env[61995]: DEBUG oslo_concurrency.lockutils [req-52e300ba-eda7-4192-8654-76f627c0879e req-3538b945-7e00-4ed3-b1f3-b593451059d4 service nova] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.822496] env[61995]: DEBUG nova.compute.manager [req-52e300ba-eda7-4192-8654-76f627c0879e req-3538b945-7e00-4ed3-b1f3-b593451059d4 service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] No waiting events found dispatching network-vif-plugged-ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.823123] env[61995]: WARNING nova.compute.manager [req-52e300ba-eda7-4192-8654-76f627c0879e req-3538b945-7e00-4ed3-b1f3-b593451059d4 service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Received unexpected event network-vif-plugged-ea813bd4-7083-4f67-b0ca-89282c951487 for instance with vm_state building and task_state block_device_mapping. [ 886.862124] env[61995]: DEBUG nova.network.neutron [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Successfully updated port: ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.895315] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100347} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.895665] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.896804] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670ac9ae-f9d6-46ae-b430-f70b1efa6309 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.925503] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5/a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.928958] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daeb769a-cd7a-4628-b12e-a609720567ec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.956779] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fdc21b-cd20-fa16-a7cf-93665cd9ed71, 'name': SearchDatastore_Task, 'duration_secs': 0.010363} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.958401] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.958687] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 886.959096] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 886.959096] env[61995]: value = "task-794742" [ 886.959096] env[61995]: _type = "Task" [ 886.959096] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.959299] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03c15bb0-aae9-485d-8153-d7fe8336d937 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.975740] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794742, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.979582] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 886.979582] env[61995]: value = "task-794743" [ 886.979582] env[61995]: _type = "Task" [ 886.979582] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.992592] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.126972] env[61995]: DEBUG nova.objects.base [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 887.154040] env[61995]: INFO nova.compute.manager [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Rebuilding instance [ 887.207765] env[61995]: DEBUG nova.compute.manager [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.208750] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0c520a-884a-4d48-a6e8-5c365435ad13 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.289378] env[61995]: DEBUG nova.network.neutron [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance_info_cache with network_info: [{"id": "3ecd14b2-d6b9-4216-936e-edae029037af", "address": "fa:16:3e:2b:38:33", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ecd14b2-d6", "ovs_interfaceid": "3ecd14b2-d6b9-4216-936e-edae029037af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.322790] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d32778-cf42-43f5-a227-dc3738985cd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.334655] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4fd26f7-9f95-4b9e-97cd-2563ebf716b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.370586] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquiring lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.370586] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquired lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.370586] env[61995]: DEBUG nova.network.neutron [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.374668] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb60af5-d844-4913-91cd-f1e590350b3d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.385142] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502fa916-be87-48c5-939b-3e6dfc3f3313 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.402436] env[61995]: DEBUG nova.compute.provider_tree [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.472023] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794742, 'name': ReconfigVM_Task, 'duration_secs': 0.492253} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.472338] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Reconfigured VM instance instance-00000045 to attach disk [datastore2] a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5/a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.472964] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-baa4ac4b-65c9-4063-92f7-2e0e428f6371 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.484300] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 887.484300] env[61995]: value = "task-794744" [ 887.484300] env[61995]: _type = "Task" [ 887.484300] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.494159] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794743, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518714} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.494833] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 887.496209] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.496209] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53418720-ef8d-4ddb-83da-5add08c7e84f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.500594] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794744, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.524366] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 887.524366] env[61995]: value = "task-794745" [ 887.524366] env[61995]: _type = "Task" [ 887.524366] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.538032] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794745, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.733074] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.734142] env[61995]: DEBUG nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.734676] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.734910] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.735018] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.735216] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.735405] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.735505] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.735739] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.735918] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.736122] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.736305] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.736761] env[61995]: DEBUG nova.virt.hardware [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.736837] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12e27ad7-bb6f-42ae-9caf-07f4c1b18b18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.739427] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18dbe41a-6b27-4995-ad8b-593f858dab11 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.749166] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed4e5b5-54a8-4a48-8e58-187f2d23d967 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.752922] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 887.752922] env[61995]: value = "task-794746" [ 887.752922] env[61995]: _type = "Task" [ 887.752922] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.771254] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 887.771464] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 887.772235] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90392966-ead0-4674-ac64-b44211315dcf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.779448] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 887.779701] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e650e4f-a646-41b1-92fd-5d2f7e7e9191 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.792891] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.857844] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 887.857844] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 887.858070] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleting the datastore file [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.858597] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2231138-6751-42e1-b709-f926517eae5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.867504] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 887.867504] env[61995]: value = "task-794748" [ 887.867504] env[61995]: _type = "Task" [ 887.867504] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.875459] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.905411] env[61995]: DEBUG nova.scheduler.client.report [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.919836] env[61995]: DEBUG nova.network.neutron [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.995744] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794744, 'name': Rename_Task, 'duration_secs': 0.170116} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.996054] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 887.996390] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d5fa251-e006-420a-934d-b8e11c38892c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.004010] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 888.004010] env[61995]: value = "task-794749" [ 888.004010] env[61995]: _type = "Task" [ 888.004010] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.015199] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794749, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.036314] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794745, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089607} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.036609] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.037539] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2193fc-a755-4d28-aa6e-7b3b7cb71c6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.041392] env[61995]: DEBUG nova.network.neutron [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.067439] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.069283] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1426e508-2058-4e69-80ab-b4e22c66a0c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.094290] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 888.094290] env[61995]: value = "task-794750" [ 888.094290] env[61995]: _type = "Task" [ 888.094290] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.106077] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794750, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.219671] env[61995]: DEBUG nova.network.neutron [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updating instance_info_cache with network_info: [{"id": "ea813bd4-7083-4f67-b0ca-89282c951487", "address": "fa:16:3e:98:89:13", "network": {"id": "88fee97d-6d56-46bf-a1da-6f7326f171fb", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-165899439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7be4efd5314088a9df13b042f7b100", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea813bd4-70", "ovs_interfaceid": "ea813bd4-7083-4f67-b0ca-89282c951487", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.326136] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc42ede-eb01-417c-97a5-d34f713b34af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.346227] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383fadee-da5f-41b1-96d3-b3e4b4e18fc2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.354255] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance '2b40609a-0826-4a16-af28-c8926af6b646' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 888.380602] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160023} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.381670] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.381670] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 888.381670] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 888.410878] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.411455] env[61995]: DEBUG nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 888.415595] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.399s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.415595] env[61995]: DEBUG nova.objects.instance [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 888.516395] env[61995]: DEBUG oslo_vmware.api [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794749, 'name': PowerOnVM_Task, 'duration_secs': 0.457031} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.516395] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 888.516395] env[61995]: INFO nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Took 9.08 seconds to spawn the instance on the hypervisor. [ 888.516395] env[61995]: DEBUG nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.517195] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfe516b-d9c1-4961-9513-33056ccb5d2c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.544070] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.605657] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794750, 'name': ReconfigVM_Task, 'duration_secs': 0.28374} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.605657] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Reconfigured VM instance instance-00000024 to attach disk [datastore2] c4f744f9-36f3-4913-9ced-f0db93d3b4df/c4f744f9-36f3-4913-9ced-f0db93d3b4df.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.606322] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15d56616-5908-420e-a445-32f566d601fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.616736] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 888.616736] env[61995]: value = "task-794751" [ 888.616736] env[61995]: _type = "Task" [ 888.616736] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.625292] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794751, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.722406] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Releasing lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.722834] env[61995]: DEBUG nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance network_info: |[{"id": "ea813bd4-7083-4f67-b0ca-89282c951487", "address": "fa:16:3e:98:89:13", "network": {"id": "88fee97d-6d56-46bf-a1da-6f7326f171fb", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-165899439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7be4efd5314088a9df13b042f7b100", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea813bd4-70", "ovs_interfaceid": "ea813bd4-7083-4f67-b0ca-89282c951487", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.723382] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:89:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea813bd4-7083-4f67-b0ca-89282c951487', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.732325] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Creating folder: Project (df7be4efd5314088a9df13b042f7b100). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 888.732724] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-caa13934-d27d-4444-9759-4852e9e5a66c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.748417] env[61995]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 888.748629] env[61995]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61995) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 888.749025] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Folder already exists: Project (df7be4efd5314088a9df13b042f7b100). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.749280] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Creating folder: Instances. Parent ref: group-v185318. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 888.749546] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87762703-a4bb-4dba-a56b-d3a71c93e289 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.761924] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Created folder: Instances in parent group-v185318. [ 888.762398] env[61995]: DEBUG oslo.service.loopingcall [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.762612] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 888.762768] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64d253c7-dc9e-4025-a1ae-9432ae237d65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.782963] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.782963] env[61995]: value = "task-794754" [ 888.782963] env[61995]: _type = "Task" [ 888.782963] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.793236] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794754, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.861636] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 888.862046] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7539275-a130-4b6e-b909-81d34789af57 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.876386] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 888.876386] env[61995]: value = "task-794755" [ 888.876386] env[61995]: _type = "Task" [ 888.876386] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.893813] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794755, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.918931] env[61995]: DEBUG nova.compute.utils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.923578] env[61995]: DEBUG nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 888.927020] env[61995]: DEBUG nova.network.neutron [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 888.964320] env[61995]: DEBUG nova.compute.manager [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Received event network-changed-ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.964616] env[61995]: DEBUG nova.compute.manager [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Refreshing instance network info cache due to event network-changed-ea813bd4-7083-4f67-b0ca-89282c951487. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 888.964909] env[61995]: DEBUG oslo_concurrency.lockutils [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] Acquiring lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.965618] env[61995]: DEBUG oslo_concurrency.lockutils [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] Acquired lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.966145] env[61995]: DEBUG nova.network.neutron [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Refreshing network info cache for port ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 889.036381] env[61995]: INFO nova.compute.manager [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Took 27.39 seconds to build instance. [ 889.049031] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.049031] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c6a1e86-7c30-4f05-b49f-fe01d2cfdbb9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.059795] env[61995]: DEBUG oslo_vmware.api [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 889.059795] env[61995]: value = "task-794756" [ 889.059795] env[61995]: _type = "Task" [ 889.059795] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.073017] env[61995]: DEBUG oslo_vmware.api [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794756, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.135528] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794751, 'name': Rename_Task, 'duration_secs': 0.15967} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.136159] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.136656] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef59c372-3ca3-4fe8-ba5a-a2d967ec7a67 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.145801] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 889.145801] env[61995]: value = "task-794758" [ 889.145801] env[61995]: _type = "Task" [ 889.145801] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.159529] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.291183] env[61995]: DEBUG nova.policy [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14d596474e4e48a4889db695a088fc72', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4de205843d14c7a91156c9b49ab7657', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 889.297387] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794754, 'name': CreateVM_Task, 'duration_secs': 0.393963} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.297670] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 889.298544] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185324', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'name': 'volume-604ce120-96fa-4557-9558-055ca5966123', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ca0dcd7-a2b8-4de6-835b-07c36ab95a99', 'attached_at': '', 'detached_at': '', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'serial': '604ce120-96fa-4557-9558-055ca5966123'}, 'attachment_id': 'e3d2a1cb-08e2-472a-89ab-1e13e2668a8c', 'guest_format': None, 'boot_index': 0, 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61995) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 889.298858] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Root volume attach. Driver type: vmdk {{(pid=61995) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 889.299994] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213090f3-2f28-4aa4-863a-2cc92ba4f8c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.308696] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329f232e-1bd9-4089-b867-d6f96e77525a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.316701] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4e7fa0-20d8-429a-85aa-ba4dd2d8046e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.324570] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-7d5ef86a-96ec-42c7-a9e3-990a44ef63c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.334158] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 889.334158] env[61995]: value = "task-794759" [ 889.334158] env[61995]: _type = "Task" [ 889.334158] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.344480] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.369203] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.369820] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.370185] env[61995]: INFO nova.compute.manager [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Shelving [ 889.388895] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794755, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.423352] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.423753] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.423997] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.424289] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.424517] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.424751] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.425085] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.425415] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.425723] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.426038] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.426345] env[61995]: DEBUG nova.virt.hardware [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.427430] env[61995]: DEBUG nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 889.432129] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1cad0d49-06e5-4efa-a737-51aa5a12641c tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.433819] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebe9449-a21c-4bc8-be39-a13c5a7e1c95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.442450] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.371s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.444567] env[61995]: INFO nova.compute.claims [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.453666] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1f47b0-3007-4ea2-a600-930fd6e71bc6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.473785] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:90:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a410e471-f28f-4398-81c2-d4cd2bc69ba8', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.483720] env[61995]: DEBUG oslo.service.loopingcall [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.485046] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 889.486314] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f770ad4f-6bee-465b-948b-11b5557a7a9e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.525674] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.525674] env[61995]: value = "task-794760" [ 889.525674] env[61995]: _type = "Task" [ 889.525674] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.534862] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794760, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.538799] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf8d6782-d12d-4ed4-a0c8-94bdc610aa3d tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.899s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.572791] env[61995]: DEBUG oslo_vmware.api [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794756, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.657023] env[61995]: DEBUG oslo_vmware.api [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794758, 'name': PowerOnVM_Task, 'duration_secs': 0.496791} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.657699] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.658076] env[61995]: DEBUG nova.compute.manager [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.658995] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ea0046-18c2-4ac4-93b1-48387fd8c8dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.848219] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 40%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.883153] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.883543] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2627e4f8-c271-4dfa-a99d-9ac131b644d8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.894581] env[61995]: DEBUG oslo_vmware.api [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794755, 'name': PowerOnVM_Task, 'duration_secs': 0.574029} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.897239] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.897705] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ac27afef-b828-40ef-bd12-5f8fb74a453e tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance '2b40609a-0826-4a16-af28-c8926af6b646' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 889.908033] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 889.908033] env[61995]: value = "task-794761" [ 889.908033] env[61995]: _type = "Task" [ 889.908033] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.920446] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794761, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.037363] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794760, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.077297] env[61995]: DEBUG oslo_vmware.api [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794756, 'name': PowerOnVM_Task} progress is 74%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.176832] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.308259] env[61995]: DEBUG nova.network.neutron [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updated VIF entry in instance network info cache for port ea813bd4-7083-4f67-b0ca-89282c951487. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 890.308680] env[61995]: DEBUG nova.network.neutron [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updating instance_info_cache with network_info: [{"id": "ea813bd4-7083-4f67-b0ca-89282c951487", "address": "fa:16:3e:98:89:13", "network": {"id": "88fee97d-6d56-46bf-a1da-6f7326f171fb", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-165899439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7be4efd5314088a9df13b042f7b100", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea813bd4-70", "ovs_interfaceid": "ea813bd4-7083-4f67-b0ca-89282c951487", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.346995] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 53%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.406218] env[61995]: DEBUG nova.network.neutron [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Successfully created port: 8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.424317] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794761, 'name': PowerOffVM_Task, 'duration_secs': 0.486463} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.425102] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.425583] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bea4ca-d68b-4155-b80f-4922a751d737 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.452428] env[61995]: DEBUG nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 890.456457] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6647651b-2c0b-4c2d-a078-b6eea805e366 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.495341] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.495626] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.495790] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.495975] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.496428] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.496619] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.496843] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.498260] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.498492] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.498680] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.498868] env[61995]: DEBUG nova.virt.hardware [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.499828] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6323feb7-4f8c-4045-8b04-29b4a705fe49 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.514361] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417365b9-7ece-47c4-b49e-ee99a3d21f72 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.544940] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794760, 'name': CreateVM_Task, 'duration_secs': 0.612786} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.545163] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.545860] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.546057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.546420] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.546695] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eddfa25-b126-4eac-9917-f9ae34a6ae1a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.554729] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 890.554729] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526be07e-8644-dc2e-02e2-808a37154d2e" [ 890.554729] env[61995]: _type = "Task" [ 890.554729] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.569467] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526be07e-8644-dc2e-02e2-808a37154d2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.578622] env[61995]: DEBUG oslo_vmware.api [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794756, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.813418] env[61995]: DEBUG oslo_concurrency.lockutils [req-391b8607-2432-4d0e-858a-3bf820f18da3 req-dea39a73-dca6-40f1-a974-9d311daea04c service nova] Releasing lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.849952] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 65%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.896104] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20a758d-250a-41f5-801b-78a9fb71c9d9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.907858] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b250a8-e9ce-4929-8efa-4cfb18dc9349 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.949112] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72834242-2c57-4847-b37a-65d222470536 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.960240] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90eb962-9938-4a6c-bac7-2a8c0c8e7188 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.984071] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 890.985464] env[61995]: DEBUG nova.compute.provider_tree [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.989047] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e0763866-58b3-472c-8927-1f19ad6e9ef2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.001698] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 891.001698] env[61995]: value = "task-794762" [ 891.001698] env[61995]: _type = "Task" [ 891.001698] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.017040] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794762, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.070823] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526be07e-8644-dc2e-02e2-808a37154d2e, 'name': SearchDatastore_Task, 'duration_secs': 0.018536} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.071676] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.071976] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.072230] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.072472] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.072615] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.072960] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5979cf9b-6d03-447c-9588-c68d43a4769e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.078474] env[61995]: DEBUG oslo_vmware.api [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794756, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.087996] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.088316] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.089171] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6a361cb-39a8-4256-bd8f-a272d8f4219c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.095689] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 891.095689] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52831e08-c63c-ca9b-7e5e-5c41acb4434a" [ 891.095689] env[61995]: _type = "Task" [ 891.095689] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.108155] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52831e08-c63c-ca9b-7e5e-5c41acb4434a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.213118] env[61995]: DEBUG nova.compute.manager [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Received event network-changed-af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 891.213249] env[61995]: DEBUG nova.compute.manager [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Refreshing instance network info cache due to event network-changed-af010e12-84c9-4a5d-b901-4f519c51ce79. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 891.213448] env[61995]: DEBUG oslo_concurrency.lockutils [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] Acquiring lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.213583] env[61995]: DEBUG oslo_concurrency.lockutils [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] Acquired lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.213751] env[61995]: DEBUG nova.network.neutron [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Refreshing network info cache for port af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.315022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.315339] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.315607] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.315928] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.316135] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.319700] env[61995]: INFO nova.compute.manager [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Terminating instance [ 891.322208] env[61995]: DEBUG nova.compute.manager [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.322506] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 891.323580] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5635633a-9ae4-4208-a688-c51b885451d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.334031] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 891.334385] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7168186-3e5c-46c4-b95d-32358fd6aa79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.344461] env[61995]: DEBUG oslo_vmware.api [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 891.344461] env[61995]: value = "task-794763" [ 891.344461] env[61995]: _type = "Task" [ 891.344461] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.353045] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 78%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.359528] env[61995]: DEBUG oslo_vmware.api [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.490734] env[61995]: DEBUG nova.scheduler.client.report [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.527855] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794762, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.575818] env[61995]: DEBUG oslo_vmware.api [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794756, 'name': PowerOnVM_Task, 'duration_secs': 2.212588} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.576242] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 891.576472] env[61995]: DEBUG nova.compute.manager [None req-7d97766e-d89e-42d6-8920-1b7f95cfdb1c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.577755] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554f37ca-aa43-4d9e-a7a8-23ee9881a6ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.607801] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52831e08-c63c-ca9b-7e5e-5c41acb4434a, 'name': SearchDatastore_Task, 'duration_secs': 0.015065} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.608817] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-031d8c6d-1e82-48ec-be0d-2928f7586687 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.615723] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 891.615723] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522233a3-af6f-8e93-e0ab-2891b547cedc" [ 891.615723] env[61995]: _type = "Task" [ 891.615723] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.625846] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522233a3-af6f-8e93-e0ab-2891b547cedc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.777818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "2b40609a-0826-4a16-af28-c8926af6b646" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.778324] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.778570] env[61995]: DEBUG nova.compute.manager [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Going to confirm migration 2 {{(pid=61995) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 891.853995] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 92%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.860220] env[61995]: DEBUG oslo_vmware.api [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794763, 'name': PowerOffVM_Task, 'duration_secs': 0.327175} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.861125] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 891.861125] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 891.861125] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f92fb152-1ba8-4738-8bcf-ce9583ce2f80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.960394] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 891.960394] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 891.960394] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleting the datastore file [datastore2] 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.960394] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b812595-deec-452c-ba95-916e37583b23 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.967748] env[61995]: DEBUG oslo_vmware.api [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 891.967748] env[61995]: value = "task-794765" [ 891.967748] env[61995]: _type = "Task" [ 891.967748] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.977735] env[61995]: DEBUG oslo_vmware.api [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.996679] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.997347] env[61995]: DEBUG nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.000395] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 10.073s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.021984] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794762, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.107842] env[61995]: DEBUG nova.network.neutron [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Updated VIF entry in instance network info cache for port af010e12-84c9-4a5d-b901-4f519c51ce79. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 892.108376] env[61995]: DEBUG nova.network.neutron [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Updating instance_info_cache with network_info: [{"id": "af010e12-84c9-4a5d-b901-4f519c51ce79", "address": "fa:16:3e:c8:6c:9a", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf010e12-84", "ovs_interfaceid": "af010e12-84c9-4a5d-b901-4f519c51ce79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.131838] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522233a3-af6f-8e93-e0ab-2891b547cedc, 'name': SearchDatastore_Task, 'duration_secs': 0.030269} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.133352] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.133804] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 892.134474] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d543e5dc-f342-4757-b180-f51546a166e6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.145094] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 892.145094] env[61995]: value = "task-794766" [ 892.145094] env[61995]: _type = "Task" [ 892.145094] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.155262] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.341282] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.341963] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.342447] env[61995]: DEBUG nova.network.neutron [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.342885] env[61995]: DEBUG nova.objects.instance [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lazy-loading 'info_cache' on Instance uuid 2b40609a-0826-4a16-af28-c8926af6b646 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.359850] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.460762] env[61995]: DEBUG nova.network.neutron [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Successfully updated port: 8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.478846] env[61995]: DEBUG oslo_vmware.api [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324432} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.479204] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.479461] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 892.479726] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 892.479944] env[61995]: INFO nova.compute.manager [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Took 1.16 seconds to destroy the instance on the hypervisor. [ 892.480259] env[61995]: DEBUG oslo.service.loopingcall [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.480546] env[61995]: DEBUG nova.compute.manager [-] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.480705] env[61995]: DEBUG nova.network.neutron [-] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 892.504073] env[61995]: DEBUG nova.compute.utils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.505759] env[61995]: DEBUG nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.506903] env[61995]: DEBUG nova.network.neutron [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 892.511371] env[61995]: INFO nova.compute.claims [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.531286] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794762, 'name': CreateSnapshot_Task, 'duration_secs': 1.195485} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.531828] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 892.533457] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6caa3558-0159-45c8-8e6a-f4e792cbf8ec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.612148] env[61995]: DEBUG oslo_concurrency.lockutils [req-ccf0ed02-e90d-4e85-9cf9-1b11816d3a67 req-2b117ecb-0aed-4459-829a-53d3cef7d09c service nova] Releasing lock "refresh_cache-a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.644357] env[61995]: DEBUG nova.policy [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caf9bbbd848d4d7ba53ba3bd553943e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2018aae65cf4bf98b5ac82519201cf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 892.673522] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794766, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.864566] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task} progress is 98%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.964402] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.964666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.964713] env[61995]: DEBUG nova.network.neutron [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.010065] env[61995]: DEBUG nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.020246] env[61995]: INFO nova.compute.resource_tracker [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating resource usage from migration 4d70102d-a1be-41ff-88ca-3e9bb403b81b [ 893.067511] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 893.070860] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ba2e2b50-0876-4f6c-81de-7704a9997069 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.085987] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 893.085987] env[61995]: value = "task-794767" [ 893.085987] env[61995]: _type = "Task" [ 893.085987] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.096785] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794767, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.163701] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.747729} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.166353] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.166738] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.170713] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb25064b-e9e6-4619-a99a-4d00ebb966e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.180899] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 893.180899] env[61995]: value = "task-794768" [ 893.180899] env[61995]: _type = "Task" [ 893.180899] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.196587] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.306655] env[61995]: DEBUG nova.network.neutron [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Successfully created port: 14cad421-d68a-4eb8-ad41-f42f7a4d644f {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.362159] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794759, 'name': RelocateVM_Task, 'duration_secs': 3.993258} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.362459] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 893.362658] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185324', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'name': 'volume-604ce120-96fa-4557-9558-055ca5966123', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ca0dcd7-a2b8-4de6-835b-07c36ab95a99', 'attached_at': '', 'detached_at': '', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'serial': '604ce120-96fa-4557-9558-055ca5966123'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 893.364233] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ef880a-eea3-4d4d-ba0a-ee25856d4565 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.393042] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b5cec4-ae74-46e9-9662-290ca20529e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.421986] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] volume-604ce120-96fa-4557-9558-055ca5966123/volume-604ce120-96fa-4557-9558-055ca5966123.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.431021] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c7291a4-7ec4-4e6f-aa16-3c4f96ba8e29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.452409] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 893.452409] env[61995]: value = "task-794769" [ 893.452409] env[61995]: _type = "Task" [ 893.452409] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.475250] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794769, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.492650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Acquiring lock "7d987006-afac-4973-ae69-f15b4752469b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.492900] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "7d987006-afac-4973-ae69-f15b4752469b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.497170] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9da9f5-b613-40af-9fa7-0e29c99a945e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.505421] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec058b79-edc4-41f9-9a8d-45001e96555d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.553169] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5720bef-163e-432f-97ca-aff39de93a39 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.563098] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fe7ba3-8177-4694-a8c3-e306c7358069 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.568727] env[61995]: DEBUG nova.network.neutron [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 893.585320] env[61995]: DEBUG nova.compute.provider_tree [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.601597] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794767, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.668097] env[61995]: DEBUG nova.network.neutron [-] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.691742] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076406} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.692019] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.692909] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1437b378-a7bd-4b12-9baa-2aa5296dd6aa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.718659] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.721543] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e60c7721-5179-41dc-9c56-5f6b35f1376c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.745729] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 893.745729] env[61995]: value = "task-794770" [ 893.745729] env[61995]: _type = "Task" [ 893.745729] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.755343] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794770, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.797189] env[61995]: DEBUG nova.network.neutron [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating instance_info_cache with network_info: [{"id": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "address": "fa:16:3e:c6:74:e1", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c280b5d-bf", "ovs_interfaceid": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.845642] env[61995]: DEBUG nova.compute.manager [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Received event network-vif-plugged-8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.845876] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.846203] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.846273] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.846498] env[61995]: DEBUG nova.compute.manager [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] No waiting events found dispatching network-vif-plugged-8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.846628] env[61995]: WARNING nova.compute.manager [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Received unexpected event network-vif-plugged-8c280b5d-bf84-4a8d-8ab7-4230645c59d1 for instance with vm_state building and task_state spawning. [ 893.846785] env[61995]: DEBUG nova.compute.manager [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Received event network-changed-8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.846961] env[61995]: DEBUG nova.compute.manager [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Refreshing instance network info cache due to event network-changed-8c280b5d-bf84-4a8d-8ab7-4230645c59d1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 893.847114] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] Acquiring lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.870620] env[61995]: DEBUG nova.network.neutron [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance_info_cache with network_info: [{"id": "3ecd14b2-d6b9-4216-936e-edae029037af", "address": "fa:16:3e:2b:38:33", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ecd14b2-d6", "ovs_interfaceid": "3ecd14b2-d6b9-4216-936e-edae029037af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.963434] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794769, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.998877] env[61995]: DEBUG nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.021112] env[61995]: DEBUG nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.054032] env[61995]: DEBUG nova.virt.hardware [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.055089] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5985132c-28f6-4b04-8c5c-8d0e703f3139 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.065434] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed26834e-bbd8-4bfc-8a15-22a9dd8a4893 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.091144] env[61995]: DEBUG nova.scheduler.client.report [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.105487] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794767, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.171452] env[61995]: INFO nova.compute.manager [-] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Took 1.69 seconds to deallocate network for instance. [ 894.259412] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.299813] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.300208] env[61995]: DEBUG nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Instance network_info: |[{"id": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "address": "fa:16:3e:c6:74:e1", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c280b5d-bf", "ovs_interfaceid": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 894.300907] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] Acquired lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.300907] env[61995]: DEBUG nova.network.neutron [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Refreshing network info cache for port 8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 894.302201] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:74:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbd7899c-c96e-47fc-9141-5803b646917a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c280b5d-bf84-4a8d-8ab7-4230645c59d1', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.312421] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Creating folder: Project (f4de205843d14c7a91156c9b49ab7657). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 894.313633] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66d92517-3a1a-487d-b552-95ca7092dfb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.325761] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Created folder: Project (f4de205843d14c7a91156c9b49ab7657) in parent group-v185203. [ 894.325965] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Creating folder: Instances. Parent ref: group-v185348. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 894.326240] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5759499-f14a-429f-ae18-becf6b307d96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.336595] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Created folder: Instances in parent group-v185348. [ 894.336855] env[61995]: DEBUG oslo.service.loopingcall [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.337062] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 894.337283] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb5ae51e-276f-43a8-a1e6-9e3aeb3b411d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.357941] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.357941] env[61995]: value = "task-794773" [ 894.357941] env[61995]: _type = "Task" [ 894.357941] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.366034] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794773, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.375772] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-2b40609a-0826-4a16-af28-c8926af6b646" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.376150] env[61995]: DEBUG nova.objects.instance [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lazy-loading 'migration_context' on Instance uuid 2b40609a-0826-4a16-af28-c8926af6b646 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.434539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-b1bd98d9-bd0f-4abd-a188-e5267ada4852-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.434539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-b1bd98d9-bd0f-4abd-a188-e5267ada4852-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.434623] env[61995]: DEBUG nova.objects.instance [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'flavor' on Instance uuid b1bd98d9-bd0f-4abd-a188-e5267ada4852 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.464047] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794769, 'name': ReconfigVM_Task, 'duration_secs': 0.845445} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.464392] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Reconfigured VM instance instance-00000046 to attach disk [datastore1] volume-604ce120-96fa-4557-9558-055ca5966123/volume-604ce120-96fa-4557-9558-055ca5966123.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.469452] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de148bf8-ad59-428b-933b-05ff631dc971 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.486236] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 894.486236] env[61995]: value = "task-794774" [ 894.486236] env[61995]: _type = "Task" [ 894.486236] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.501346] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794774, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.523890] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.600010] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.600s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.600422] env[61995]: INFO nova.compute.manager [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Migrating [ 894.606939] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794767, 'name': CloneVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.607246] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.096s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.608880] env[61995]: INFO nova.compute.claims [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.679925] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.757067] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794770, 'name': ReconfigVM_Task, 'duration_secs': 0.756885} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.757482] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Reconfigured VM instance instance-00000044 to attach disk [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff/fdeceddf-c6f9-4a98-afd0-defa1284eeff.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.758205] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5a233c0-a77a-4d8d-bf92-10f6038a31c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.765545] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 894.765545] env[61995]: value = "task-794775" [ 894.765545] env[61995]: _type = "Task" [ 894.765545] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.776793] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794775, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.867406] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794773, 'name': CreateVM_Task, 'duration_secs': 0.43452} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.867572] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 894.868302] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.868474] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.868795] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.869082] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76ef2f83-571b-4f5f-a0c7-fc43a1a4658b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.874305] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 894.874305] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e9968f-3ea1-d125-abce-358dc0c4e483" [ 894.874305] env[61995]: _type = "Task" [ 894.874305] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.878640] env[61995]: DEBUG nova.objects.base [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Object Instance<2b40609a-0826-4a16-af28-c8926af6b646> lazy-loaded attributes: info_cache,migration_context {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 894.879461] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e96d3b9-0507-4b58-8d7e-7db4949868d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.887678] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e9968f-3ea1-d125-abce-358dc0c4e483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.906902] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87bbcba0-8a13-42e0-9f45-e8f52ae6986b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.913252] env[61995]: DEBUG oslo_vmware.api [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 894.913252] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522e8d61-db41-c387-72a2-a36a3dbd2d90" [ 894.913252] env[61995]: _type = "Task" [ 894.913252] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.925992] env[61995]: DEBUG oslo_vmware.api [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522e8d61-db41-c387-72a2-a36a3dbd2d90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.997365] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794774, 'name': ReconfigVM_Task, 'duration_secs': 0.193093} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.997502] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185324', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'name': 'volume-604ce120-96fa-4557-9558-055ca5966123', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ca0dcd7-a2b8-4de6-835b-07c36ab95a99', 'attached_at': '', 'detached_at': '', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'serial': '604ce120-96fa-4557-9558-055ca5966123'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 894.998041] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-759bd90f-2137-4b07-809d-7e751edf6eb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.005752] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 895.005752] env[61995]: value = "task-794776" [ 895.005752] env[61995]: _type = "Task" [ 895.005752] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.015278] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794776, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.068752] env[61995]: DEBUG nova.objects.instance [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'pci_requests' on Instance uuid b1bd98d9-bd0f-4abd-a188-e5267ada4852 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.101930] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794767, 'name': CloneVM_Task, 'duration_secs': 1.583718} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.102368] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Created linked-clone VM from snapshot [ 895.103546] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215ecf02-35e6-4b29-8c91-980778bf3fef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.113515] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Uploading image d9aa4058-3449-47b3-b10b-42d75d2fcc13 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 895.121737] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.121937] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.122148] env[61995]: DEBUG nova.network.neutron [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 895.143817] env[61995]: DEBUG nova.network.neutron [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updated VIF entry in instance network info cache for port 8c280b5d-bf84-4a8d-8ab7-4230645c59d1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.144240] env[61995]: DEBUG nova.network.neutron [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating instance_info_cache with network_info: [{"id": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "address": "fa:16:3e:c6:74:e1", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c280b5d-bf", "ovs_interfaceid": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.152230] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 895.152230] env[61995]: value = "vm-185347" [ 895.152230] env[61995]: _type = "VirtualMachine" [ 895.152230] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 895.152483] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1ebdca43-fca3-4b11-9826-d99ce17d88ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.161782] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lease: (returnval){ [ 895.161782] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52de7cab-e768-abc4-fd91-716861a1a62e" [ 895.161782] env[61995]: _type = "HttpNfcLease" [ 895.161782] env[61995]: } obtained for exporting VM: (result){ [ 895.161782] env[61995]: value = "vm-185347" [ 895.161782] env[61995]: _type = "VirtualMachine" [ 895.161782] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 895.162139] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the lease: (returnval){ [ 895.162139] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52de7cab-e768-abc4-fd91-716861a1a62e" [ 895.162139] env[61995]: _type = "HttpNfcLease" [ 895.162139] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 895.170164] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 895.170164] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52de7cab-e768-abc4-fd91-716861a1a62e" [ 895.170164] env[61995]: _type = "HttpNfcLease" [ 895.170164] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 895.276295] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794775, 'name': Rename_Task, 'duration_secs': 0.33228} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.276573] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.276822] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5b8439f-c808-4e42-8125-3add454650ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.283814] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 895.283814] env[61995]: value = "task-794778" [ 895.283814] env[61995]: _type = "Task" [ 895.283814] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.291872] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794778, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.384878] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e9968f-3ea1-d125-abce-358dc0c4e483, 'name': SearchDatastore_Task, 'duration_secs': 0.016885} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.385297] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.385541] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.385780] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.385928] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.386143] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.386407] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f63cc4f2-99ba-4225-9b42-0a5b1d89fecb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.394845] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.395036] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 895.395840] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-749ba994-c7a0-416c-9f31-d9ff1e9691cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.401485] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 895.401485] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52138def-0d2e-d7f4-3ea2-a0796e616a93" [ 895.401485] env[61995]: _type = "Task" [ 895.401485] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.411567] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52138def-0d2e-d7f4-3ea2-a0796e616a93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.422218] env[61995]: DEBUG oslo_vmware.api [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522e8d61-db41-c387-72a2-a36a3dbd2d90, 'name': SearchDatastore_Task, 'duration_secs': 0.01016} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.422489] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.525272] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794776, 'name': Rename_Task, 'duration_secs': 0.2055} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.525588] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.525837] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d20adb71-aab5-4512-ad5c-36b3f8c6c883 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.533224] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 895.533224] env[61995]: value = "task-794779" [ 895.533224] env[61995]: _type = "Task" [ 895.533224] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.542322] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794779, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.549083] env[61995]: DEBUG nova.compute.manager [req-1955a4c9-fd76-4a45-a65c-ef8254925e60 req-5df8d39d-9c8c-4e81-952c-2fa5fe849242 service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Received event network-vif-plugged-14cad421-d68a-4eb8-ad41-f42f7a4d644f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.549260] env[61995]: DEBUG oslo_concurrency.lockutils [req-1955a4c9-fd76-4a45-a65c-ef8254925e60 req-5df8d39d-9c8c-4e81-952c-2fa5fe849242 service nova] Acquiring lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.549505] env[61995]: DEBUG oslo_concurrency.lockutils [req-1955a4c9-fd76-4a45-a65c-ef8254925e60 req-5df8d39d-9c8c-4e81-952c-2fa5fe849242 service nova] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.549646] env[61995]: DEBUG oslo_concurrency.lockutils [req-1955a4c9-fd76-4a45-a65c-ef8254925e60 req-5df8d39d-9c8c-4e81-952c-2fa5fe849242 service nova] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.549875] env[61995]: DEBUG nova.compute.manager [req-1955a4c9-fd76-4a45-a65c-ef8254925e60 req-5df8d39d-9c8c-4e81-952c-2fa5fe849242 service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] No waiting events found dispatching network-vif-plugged-14cad421-d68a-4eb8-ad41-f42f7a4d644f {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.549984] env[61995]: WARNING nova.compute.manager [req-1955a4c9-fd76-4a45-a65c-ef8254925e60 req-5df8d39d-9c8c-4e81-952c-2fa5fe849242 service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Received unexpected event network-vif-plugged-14cad421-d68a-4eb8-ad41-f42f7a4d644f for instance with vm_state building and task_state spawning. [ 895.571535] env[61995]: DEBUG nova.objects.base [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 895.571762] env[61995]: DEBUG nova.network.neutron [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 895.612309] env[61995]: DEBUG nova.policy [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 895.646797] env[61995]: DEBUG oslo_concurrency.lockutils [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] Releasing lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.647151] env[61995]: DEBUG nova.compute.manager [req-9e90d751-7ea6-4dc5-8949-948e690116fc req-fc7a5784-825b-42ed-9e7e-c1a8de225930 service nova] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Received event network-vif-deleted-b25c9bc9-03d6-4bcd-a3e6-8469128cea27 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.677449] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 895.677449] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52de7cab-e768-abc4-fd91-716861a1a62e" [ 895.677449] env[61995]: _type = "HttpNfcLease" [ 895.677449] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 895.677974] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 895.677974] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52de7cab-e768-abc4-fd91-716861a1a62e" [ 895.677974] env[61995]: _type = "HttpNfcLease" [ 895.677974] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 895.681638] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babe627c-bc49-479a-8354-dae975ca5873 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.692558] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6e39a-3c4d-4284-bd0a-59d993d2716e/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 895.692797] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6e39a-3c4d-4284-bd0a-59d993d2716e/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 895.754117] env[61995]: DEBUG nova.network.neutron [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Successfully updated port: 14cad421-d68a-4eb8-ad41-f42f7a4d644f {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.787654] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7428b92c-9805-434c-a8b3-8f5c49153b22 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.800583] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794778, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.915324] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52138def-0d2e-d7f4-3ea2-a0796e616a93, 'name': SearchDatastore_Task, 'duration_secs': 0.014753} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.915324] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-056ebc79-cd15-41f2-b3e9-3b6158a43f38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.926820] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 895.926820] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b404a5-45b7-2928-5417-c3d51251dc05" [ 895.926820] env[61995]: _type = "Task" [ 895.926820] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.941277] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b404a5-45b7-2928-5417-c3d51251dc05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.945093] env[61995]: DEBUG nova.network.neutron [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Successfully created port: 7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 895.996061] env[61995]: DEBUG nova.network.neutron [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance_info_cache with network_info: [{"id": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "address": "fa:16:3e:45:26:f9", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb488f7-74", "ovs_interfaceid": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.046715] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794779, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.061586] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfa2513-fa06-4120-b9bd-ed42c5aa54e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.071949] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b66aae1-deea-44cc-901a-9d21ab535e0b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.113932] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f5a183-303a-42ee-928e-a1426976cedc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.124372] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9270811-ea54-4bc4-845b-efa207f74bd5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.144852] env[61995]: DEBUG nova.compute.provider_tree [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.193102] env[61995]: DEBUG nova.compute.manager [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Received event network-changed-14cad421-d68a-4eb8-ad41-f42f7a4d644f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.193475] env[61995]: DEBUG nova.compute.manager [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Refreshing instance network info cache due to event network-changed-14cad421-d68a-4eb8-ad41-f42f7a4d644f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 896.193808] env[61995]: DEBUG oslo_concurrency.lockutils [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] Acquiring lock "refresh_cache-a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.194098] env[61995]: DEBUG oslo_concurrency.lockutils [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] Acquired lock "refresh_cache-a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.194373] env[61995]: DEBUG nova.network.neutron [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Refreshing network info cache for port 14cad421-d68a-4eb8-ad41-f42f7a4d644f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.257671] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "refresh_cache-a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.295266] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794778, 'name': PowerOnVM_Task, 'duration_secs': 0.871444} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.295560] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.295934] env[61995]: DEBUG nova.compute.manager [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.296956] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab7cb0b-2b8c-4f0f-8196-20947a98c7c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.438257] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b404a5-45b7-2928-5417-c3d51251dc05, 'name': SearchDatastore_Task, 'duration_secs': 0.011601} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.438706] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.439073] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] e835ac53-29af-4bd0-b186-5c6270ccf760/e835ac53-29af-4bd0-b186-5c6270ccf760.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 896.439390] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49e8c51c-efd3-43ce-a2ec-56e6384990d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.449126] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 896.449126] env[61995]: value = "task-794780" [ 896.449126] env[61995]: _type = "Task" [ 896.449126] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.460133] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794780, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.499945] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.545783] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794779, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.647835] env[61995]: DEBUG nova.scheduler.client.report [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.734638] env[61995]: DEBUG nova.network.neutron [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 896.809812] env[61995]: INFO nova.compute.manager [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] bringing vm to original state: 'stopped' [ 896.865431] env[61995]: DEBUG nova.network.neutron [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.963814] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794780, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.048216] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794779, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.152949] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.153649] env[61995]: DEBUG nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.157177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.980s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.157177] env[61995]: DEBUG nova.objects.instance [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 897.369460] env[61995]: DEBUG oslo_concurrency.lockutils [req-82d951ac-c735-4393-9518-1b3e0e137480 req-acc6d89c-6b8c-49d6-94d8-b3c90e48838f service nova] Releasing lock "refresh_cache-a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.370130] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "refresh_cache-a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.370938] env[61995]: DEBUG nova.network.neutron [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.461640] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794780, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.623348} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.461707] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] e835ac53-29af-4bd0-b186-5c6270ccf760/e835ac53-29af-4bd0-b186-5c6270ccf760.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 897.462164] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.462535] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6ec7923-5fab-475b-a0c6-0ce395607058 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.471777] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 897.471777] env[61995]: value = "task-794781" [ 897.471777] env[61995]: _type = "Task" [ 897.471777] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.481805] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794781, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.501673] env[61995]: DEBUG nova.network.neutron [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Successfully updated port: 7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.549228] env[61995]: DEBUG oslo_vmware.api [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794779, 'name': PowerOnVM_Task, 'duration_secs': 1.954083} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.550046] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 897.550136] env[61995]: INFO nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Took 9.82 seconds to spawn the instance on the hypervisor. [ 897.550338] env[61995]: DEBUG nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.551362] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6621116a-f71a-47f1-9f50-d93495ad60c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.662752] env[61995]: DEBUG nova.compute.utils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.667864] env[61995]: DEBUG nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.668432] env[61995]: DEBUG nova.network.neutron [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 897.710792] env[61995]: DEBUG nova.policy [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '875241264c4b48119d0bf06a53bad1bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '837124ca1f354c8aae242a8af95756a7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 897.820551] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.820884] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.821087] env[61995]: DEBUG nova.compute.manager [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.822068] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7c6e91-aaaa-4098-ac1f-9166a0295d1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.830427] env[61995]: DEBUG nova.compute.manager [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 897.832983] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 897.833315] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-058f1bb4-b4be-41d0-b40a-4f7b2e234b81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.842181] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 897.842181] env[61995]: value = "task-794782" [ 897.842181] env[61995]: _type = "Task" [ 897.842181] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.855372] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794782, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.903294] env[61995]: DEBUG nova.network.neutron [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 898.800086] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.800507] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.800556] env[61995]: DEBUG nova.network.neutron [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.808738] env[61995]: DEBUG nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.815513] env[61995]: DEBUG oslo_concurrency.lockutils [None req-733c447c-3700-4444-87dd-27461fb05128 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.656s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.815513] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794781, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137276} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.818355] env[61995]: DEBUG nova.network.neutron [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Successfully created port: 05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.826756] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.303s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.828656] env[61995]: INFO nova.compute.claims [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.833537] env[61995]: DEBUG nova.compute.manager [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received event network-vif-plugged-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.833537] env[61995]: DEBUG oslo_concurrency.lockutils [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.833690] env[61995]: DEBUG oslo_concurrency.lockutils [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.833900] env[61995]: DEBUG oslo_concurrency.lockutils [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.834153] env[61995]: DEBUG nova.compute.manager [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] No waiting events found dispatching network-vif-plugged-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.834416] env[61995]: WARNING nova.compute.manager [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received unexpected event network-vif-plugged-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 for instance with vm_state active and task_state None. [ 898.834653] env[61995]: DEBUG nova.compute.manager [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received event network-changed-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.834871] env[61995]: DEBUG nova.compute.manager [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Refreshing instance network info cache due to event network-changed-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 898.835115] env[61995]: DEBUG oslo_concurrency.lockutils [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] Acquiring lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.839027] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.840203] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c96835-83c1-472d-a32a-09aa4a37a115 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.843284] env[61995]: INFO nova.compute.manager [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Took 25.48 seconds to build instance. [ 898.845747] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331f73cc-cc67-4b60-b52d-330a6e22bf6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.858927] env[61995]: DEBUG oslo_vmware.api [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794782, 'name': PowerOffVM_Task, 'duration_secs': 0.320694} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.882612] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 898.882966] env[61995]: DEBUG nova.compute.manager [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.884535] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance '4b49e7b1-7fd0-4e59-af79-5c0898967c35' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 898.896616] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] e835ac53-29af-4bd0-b186-5c6270ccf760/e835ac53-29af-4bd0-b186-5c6270ccf760.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.898557] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ca25d1-63e4-4c3b-90f2-de43c7245c39 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.901523] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff8597d2-5cb9-4d1d-bdc1-14bc1c1fe4fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.919951] env[61995]: WARNING nova.network.neutron [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 898.935901] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 898.935901] env[61995]: value = "task-794783" [ 898.935901] env[61995]: _type = "Task" [ 898.935901] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.946347] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794783, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.000801] env[61995]: DEBUG nova.network.neutron [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Updating instance_info_cache with network_info: [{"id": "14cad421-d68a-4eb8-ad41-f42f7a4d644f", "address": "fa:16:3e:91:59:56", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14cad421-d6", "ovs_interfaceid": "14cad421-d68a-4eb8-ad41-f42f7a4d644f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.351523] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f6a05649-a55c-4a0e-bfd5-4e19f3646856 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.997s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.400153] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.400493] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fb65964-242e-469a-ba8b-8982fe76a279 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.412863] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 899.412863] env[61995]: value = "task-794784" [ 899.412863] env[61995]: _type = "Task" [ 899.412863] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.427266] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.429075] env[61995]: DEBUG nova.network.neutron [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7812f1a4-43be-4dc0-82a1-b92cd78fb1e1", "address": "fa:16:3e:34:f8:40", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7812f1a4-43", "ovs_interfaceid": "7812f1a4-43be-4dc0-82a1-b92cd78fb1e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.443556] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.622s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.451973] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794783, 'name': ReconfigVM_Task, 'duration_secs': 0.419007} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.452536] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Reconfigured VM instance instance-00000047 to attach disk [datastore1] e835ac53-29af-4bd0-b186-5c6270ccf760/e835ac53-29af-4bd0-b186-5c6270ccf760.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.453648] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9472f384-2642-4023-a63f-61a7c3c75350 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.462148] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 899.462148] env[61995]: value = "task-794785" [ 899.462148] env[61995]: _type = "Task" [ 899.462148] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.472395] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794785, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.503415] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "refresh_cache-a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.503850] env[61995]: DEBUG nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Instance network_info: |[{"id": "14cad421-d68a-4eb8-ad41-f42f7a4d644f", "address": "fa:16:3e:91:59:56", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14cad421-d6", "ovs_interfaceid": "14cad421-d68a-4eb8-ad41-f42f7a4d644f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.504419] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:59:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14cad421-d68a-4eb8-ad41-f42f7a4d644f', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.512881] env[61995]: DEBUG oslo.service.loopingcall [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.513205] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 899.513481] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b1a4e7a-1c75-4b4d-b44f-a79a33fc25b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.536084] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.536084] env[61995]: value = "task-794786" [ 899.536084] env[61995]: _type = "Task" [ 899.536084] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.545892] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794786, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.827225] env[61995]: DEBUG nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.854082] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.854427] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.854657] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.854997] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.855103] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.855295] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.855556] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.855757] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.855989] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.856223] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.856458] env[61995]: DEBUG nova.virt.hardware [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.857472] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b050147-872a-4367-a0b1-c1db162bf54e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.867081] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83085607-7393-46f1-adb1-df24a709ac41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.925383] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794784, 'name': PowerOffVM_Task, 'duration_secs': 0.185827} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.928212] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 899.928408] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance '4b49e7b1-7fd0-4e59-af79-5c0898967c35' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 899.932240] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.932887] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.933057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.933574] env[61995]: DEBUG oslo_concurrency.lockutils [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] Acquired lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.933767] env[61995]: DEBUG nova.network.neutron [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Refreshing network info cache for port 7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 899.935973] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228395d5-1ee8-4ec3-bf04-bc0f447ebdb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.956967] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.956967] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.956967] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.957206] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.957282] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.957470] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.957763] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.957952] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.958199] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.958443] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.958663] env[61995]: DEBUG nova.virt.hardware [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.965245] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Reconfiguring VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 899.970763] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eeda244b-7ebf-4fb3-ac4b-c8f6d602311f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.984743] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.994586] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794785, 'name': Rename_Task, 'duration_secs': 0.195545} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.996259] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.996507] env[61995]: DEBUG oslo_vmware.api [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 899.996507] env[61995]: value = "task-794787" [ 899.996507] env[61995]: _type = "Task" [ 899.996507] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.999181] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56d115c9-526d-4d48-8812-660b6a8e429e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.012386] env[61995]: DEBUG oslo_vmware.api [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794787, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.013922] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 900.013922] env[61995]: value = "task-794788" [ 900.013922] env[61995]: _type = "Task" [ 900.013922] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.024915] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794788, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.046726] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794786, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.233442] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2345b3-c18d-4f35-bfb1-99b76ad791c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.242969] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35eec9a3-d1f9-4624-997a-127bb4a0b433 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.276733] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdb8cfe-38b4-4099-a7c4-f2e9060a3543 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.285915] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0941670-dcd6-46ac-bb0b-a483bff150db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.301122] env[61995]: DEBUG nova.compute.provider_tree [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.436025] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.436376] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.436541] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.436736] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.436885] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.437057] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.437651] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.437651] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.437651] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.437863] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.437939] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.445693] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cbf80f4-34ff-466f-8f48-9d6b70d322a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.466040] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 900.466040] env[61995]: value = "task-794789" [ 900.466040] env[61995]: _type = "Task" [ 900.466040] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.477115] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794789, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.511129] env[61995]: DEBUG oslo_vmware.api [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794787, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.529907] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794788, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.552288] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794786, 'name': CreateVM_Task, 'duration_secs': 0.591707} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.552515] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 900.553270] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.553498] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.553870] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.554509] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47734899-ed68-4d9b-bfb5-66717ad2c9e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.559376] env[61995]: DEBUG nova.compute.manager [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Received event network-changed-ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 900.560237] env[61995]: DEBUG nova.compute.manager [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Refreshing instance network info cache due to event network-changed-ea813bd4-7083-4f67-b0ca-89282c951487. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 900.560605] env[61995]: DEBUG oslo_concurrency.lockutils [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] Acquiring lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.560776] env[61995]: DEBUG oslo_concurrency.lockutils [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] Acquired lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.560973] env[61995]: DEBUG nova.network.neutron [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Refreshing network info cache for port ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.566022] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 900.566022] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f17060-b32f-7fde-ee17-41dc7be1214d" [ 900.566022] env[61995]: _type = "Task" [ 900.566022] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.575981] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f17060-b32f-7fde-ee17-41dc7be1214d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.804577] env[61995]: DEBUG nova.scheduler.client.report [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.838725] env[61995]: DEBUG nova.network.neutron [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updated VIF entry in instance network info cache for port 7812f1a4-43be-4dc0-82a1-b92cd78fb1e1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 900.839453] env[61995]: DEBUG nova.network.neutron [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7812f1a4-43be-4dc0-82a1-b92cd78fb1e1", "address": "fa:16:3e:34:f8:40", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7812f1a4-43", "ovs_interfaceid": "7812f1a4-43be-4dc0-82a1-b92cd78fb1e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.886274] env[61995]: DEBUG nova.network.neutron [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Successfully updated port: 05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.979707] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794789, 'name': ReconfigVM_Task, 'duration_secs': 0.214626} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.979971] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance '4b49e7b1-7fd0-4e59-af79-5c0898967c35' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 901.013885] env[61995]: DEBUG oslo_vmware.api [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794787, 'name': ReconfigVM_Task, 'duration_secs': 0.885077} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.014690] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.015018] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Reconfigured VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 901.019658] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.019898] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.020162] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.020430] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.021043] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.026084] env[61995]: INFO nova.compute.manager [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Terminating instance [ 901.028325] env[61995]: DEBUG nova.compute.manager [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.028558] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 901.029839] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2bc710-9953-4489-b91e-1f0abe267fa1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.036076] env[61995]: DEBUG oslo_vmware.api [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794788, 'name': PowerOnVM_Task, 'duration_secs': 0.781023} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.037486] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 901.037747] env[61995]: INFO nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Took 10.59 seconds to spawn the instance on the hypervisor. [ 901.037956] env[61995]: DEBUG nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.039437] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8b2ba6-1599-497e-afb7-1f90e418dce4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.044531] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 901.045300] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0737a0d-76c7-4a8b-aea3-57bc25b8c604 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.084097] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f17060-b32f-7fde-ee17-41dc7be1214d, 'name': SearchDatastore_Task, 'duration_secs': 0.012656} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.084097] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.084650] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.085514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.085514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.085514] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.086962] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d749869-e45e-40a7-bf8a-03e2482240d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.097122] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.097454] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 901.098293] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69a8a1ef-7f2d-4773-8756-9f2677da0e40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.105736] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 901.105736] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c34d91-8a53-f2fd-d1c4-c38e545e18ae" [ 901.105736] env[61995]: _type = "Task" [ 901.105736] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.121419] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c34d91-8a53-f2fd-d1c4-c38e545e18ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.128332] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 901.128680] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 901.129239] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleting the datastore file [datastore1] fdeceddf-c6f9-4a98-afd0-defa1284eeff {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.129728] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00649be6-cc24-4efb-b05d-04a85b16190b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.140870] env[61995]: DEBUG oslo_vmware.api [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 901.140870] env[61995]: value = "task-794791" [ 901.140870] env[61995]: _type = "Task" [ 901.140870] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.156582] env[61995]: DEBUG oslo_vmware.api [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.309816] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.310092] env[61995]: DEBUG nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 901.312923] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.633s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.313172] env[61995]: DEBUG nova.objects.instance [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lazy-loading 'resources' on Instance uuid 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.323054] env[61995]: DEBUG nova.network.neutron [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updated VIF entry in instance network info cache for port ea813bd4-7083-4f67-b0ca-89282c951487. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 901.323054] env[61995]: DEBUG nova.network.neutron [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updating instance_info_cache with network_info: [{"id": "ea813bd4-7083-4f67-b0ca-89282c951487", "address": "fa:16:3e:98:89:13", "network": {"id": "88fee97d-6d56-46bf-a1da-6f7326f171fb", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-165899439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df7be4efd5314088a9df13b042f7b100", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea813bd4-70", "ovs_interfaceid": "ea813bd4-7083-4f67-b0ca-89282c951487", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.342447] env[61995]: DEBUG oslo_concurrency.lockutils [req-68012e1f-af25-4ca5-b933-929bb5f45a2e req-3c953c0a-6c37-46ac-b1d0-319dd50d5698 service nova] Releasing lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.388336] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.388499] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquired lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.388687] env[61995]: DEBUG nova.network.neutron [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.488164] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:58:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2ec0e7a2-ad31-4b6b-b0cd-3d774a987d95',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-892104472',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.488459] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.488640] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.488833] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.488984] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.489380] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.489645] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.489869] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.490096] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.490378] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.490624] env[61995]: DEBUG nova.virt.hardware [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.496000] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfiguring VM instance instance-00000043 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 901.496429] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b4cdfa5-bc57-499f-ac15-98a894f618ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.521485] env[61995]: DEBUG oslo_concurrency.lockutils [None req-29a1b8a8-a9ac-4d89-8ba0-ce1912cb2bc2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-b1bd98d9-bd0f-4abd-a188-e5267ada4852-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.087s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.525051] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 901.525051] env[61995]: value = "task-794792" [ 901.525051] env[61995]: _type = "Task" [ 901.525051] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.535643] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794792, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.561314] env[61995]: INFO nova.compute.manager [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Took 25.65 seconds to build instance. [ 901.616859] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c34d91-8a53-f2fd-d1c4-c38e545e18ae, 'name': SearchDatastore_Task, 'duration_secs': 0.018782} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.617695] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cc68ed4-8a4d-444b-9513-a2fca1bed8d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.623869] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 901.623869] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527e3c9c-6173-ae2b-47c6-809c1b148254" [ 901.623869] env[61995]: _type = "Task" [ 901.623869] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.632973] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527e3c9c-6173-ae2b-47c6-809c1b148254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.650729] env[61995]: DEBUG oslo_vmware.api [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794791, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.434783} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.651027] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.651294] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 901.651529] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 901.651681] env[61995]: INFO nova.compute.manager [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Took 0.62 seconds to destroy the instance on the hypervisor. [ 901.651953] env[61995]: DEBUG oslo.service.loopingcall [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.652222] env[61995]: DEBUG nova.compute.manager [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.652410] env[61995]: DEBUG nova.network.neutron [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 901.826041] env[61995]: DEBUG nova.compute.utils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.828563] env[61995]: DEBUG oslo_concurrency.lockutils [req-2720e0f3-8e15-4b4d-8a9d-52bc0247301b req-028eb842-3971-45f9-bc2d-4909f522487b service nova] Releasing lock "refresh_cache-4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.829176] env[61995]: DEBUG nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.829451] env[61995]: DEBUG nova.network.neutron [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 901.890310] env[61995]: DEBUG nova.policy [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd2110033c4f4624afb98c6888f94de3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1ed911ecc2748cc9ec398029d3b3e78', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 901.977607] env[61995]: DEBUG nova.network.neutron [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 902.049215] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794792, 'name': ReconfigVM_Task, 'duration_secs': 0.192065} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.049786] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfigured VM instance instance-00000043 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 902.050837] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229ba1c6-7f7d-4e02-8f12-a9cec4c6c90e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.079284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4fecee7e-325e-4613-b555-91318939c40a tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.183s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.087571] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.093705] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54b8eca5-4876-49eb-92a3-7fe900cef2da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.116999] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 902.116999] env[61995]: value = "task-794793" [ 902.116999] env[61995]: _type = "Task" [ 902.116999] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.137032] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794793, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.144011] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527e3c9c-6173-ae2b-47c6-809c1b148254, 'name': SearchDatastore_Task, 'duration_secs': 0.015569} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.144011] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.144334] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] a8b7bcd6-4753-4e1f-8b0b-777c882b99f5/a8b7bcd6-4753-4e1f-8b0b-777c882b99f5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 902.144621] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c5e61d4-f8f6-4a3a-98b8-ab40b952f490 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.155918] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 902.155918] env[61995]: value = "task-794794" [ 902.155918] env[61995]: _type = "Task" [ 902.155918] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.166932] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.269722] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797b6b9d-e139-48ba-ba4b-a4bdd62a5a11 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.274713] env[61995]: DEBUG nova.network.neutron [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Successfully created port: a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.283075] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1606d5ea-d0ab-43c7-b021-d0d4c829bece {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.321925] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243455ac-565b-4655-abf7-79aadd0c0674 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.331041] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631d0631-062f-4b9a-aeec-fd30d714dfb3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.335803] env[61995]: DEBUG nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 902.351942] env[61995]: DEBUG nova.compute.provider_tree [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.443290] env[61995]: DEBUG nova.network.neutron [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Updating instance_info_cache with network_info: [{"id": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "address": "fa:16:3e:c9:4f:e8", "network": {"id": "3f7f34b5-aa6b-4e83-8593-e51a8f0544e8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-50477412-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "837124ca1f354c8aae242a8af95756a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a69ed1dd-213a-4e30-992a-466735188bf6", "external-id": "nsx-vlan-transportzone-102", "segmentation_id": 102, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05b0b88f-30", "ovs_interfaceid": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.520295] env[61995]: DEBUG nova.network.neutron [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.636751] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794793, 'name': ReconfigVM_Task, 'duration_secs': 0.291297} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.637089] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.637389] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance '4b49e7b1-7fd0-4e59-af79-5c0898967c35' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 902.669420] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794794, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.842618] env[61995]: INFO nova.virt.block_device [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Booting with volume e9de6b68-7402-4a32-b694-1d36fc3a9a02 at /dev/sda [ 902.845755] env[61995]: DEBUG nova.compute.manager [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Received event network-vif-plugged-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.846023] env[61995]: DEBUG oslo_concurrency.lockutils [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] Acquiring lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.846225] env[61995]: DEBUG oslo_concurrency.lockutils [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.846398] env[61995]: DEBUG oslo_concurrency.lockutils [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.846566] env[61995]: DEBUG nova.compute.manager [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] No waiting events found dispatching network-vif-plugged-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.846729] env[61995]: WARNING nova.compute.manager [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Received unexpected event network-vif-plugged-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f for instance with vm_state building and task_state spawning. [ 902.846973] env[61995]: DEBUG nova.compute.manager [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Received event network-changed-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.847055] env[61995]: DEBUG nova.compute.manager [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Refreshing instance network info cache due to event network-changed-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 902.847218] env[61995]: DEBUG oslo_concurrency.lockutils [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] Acquiring lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.855428] env[61995]: DEBUG nova.scheduler.client.report [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.884718] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-13cc73ba-2b80-4c35-9f6e-140328fc66d8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.896325] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916639cc-d967-4cfb-a8a6-90f125e9ebe4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.920097] env[61995]: DEBUG nova.compute.manager [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Received event network-changed-8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.920330] env[61995]: DEBUG nova.compute.manager [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Refreshing instance network info cache due to event network-changed-8c280b5d-bf84-4a8d-8ab7-4230645c59d1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 902.920537] env[61995]: DEBUG oslo_concurrency.lockutils [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] Acquiring lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.920691] env[61995]: DEBUG oslo_concurrency.lockutils [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] Acquired lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.920857] env[61995]: DEBUG nova.network.neutron [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Refreshing network info cache for port 8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.935356] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-084ca24d-f380-45f8-8a63-a23a882bbd14 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.947158] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9168158c-29a0-489e-8939-4fb4a985438d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.958359] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Releasing lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.958670] env[61995]: DEBUG nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Instance network_info: |[{"id": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "address": "fa:16:3e:c9:4f:e8", "network": {"id": "3f7f34b5-aa6b-4e83-8593-e51a8f0544e8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-50477412-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "837124ca1f354c8aae242a8af95756a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a69ed1dd-213a-4e30-992a-466735188bf6", "external-id": "nsx-vlan-transportzone-102", "segmentation_id": 102, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05b0b88f-30", "ovs_interfaceid": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.959241] env[61995]: DEBUG oslo_concurrency.lockutils [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] Acquired lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.959426] env[61995]: DEBUG nova.network.neutron [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Refreshing network info cache for port 05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.960616] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:4f:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a69ed1dd-213a-4e30-992a-466735188bf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.969048] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Creating folder: Project (837124ca1f354c8aae242a8af95756a7). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 902.970934] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa16b56c-777a-4c07-a81e-14db72249319 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.000103] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68a30df-e5d7-4de0-8079-40ac5c8e43bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.002937] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Created folder: Project (837124ca1f354c8aae242a8af95756a7) in parent group-v185203. [ 903.003145] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Creating folder: Instances. Parent ref: group-v185352. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 903.003382] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b0a9d05-40da-4734-a622-283367a339e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.010556] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fcedcd-4da6-424e-9513-0f745859b8a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.014861] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Created folder: Instances in parent group-v185352. [ 903.015124] env[61995]: DEBUG oslo.service.loopingcall [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.015322] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 903.016385] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9437835-cb8b-4030-aaf2-b3cccae9ce95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.034461] env[61995]: INFO nova.compute.manager [-] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Took 1.38 seconds to deallocate network for instance. [ 903.034849] env[61995]: DEBUG nova.virt.block_device [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updating existing volume attachment record: 61bf43d0-0bd1-4d9d-80e9-33372f9da32c {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 903.046659] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.046659] env[61995]: value = "task-794797" [ 903.046659] env[61995]: _type = "Task" [ 903.046659] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.053908] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794797, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.149426] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05809b0e-969a-4a42-8f0b-a4502c665094 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.177601] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8239acf7-28a2-4be9-aa49-f15cb0574064 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.183928] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794794, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.200227] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance '4b49e7b1-7fd0-4e59-af79-5c0898967c35' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 903.361763] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.049s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.364292] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.942s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.392472] env[61995]: INFO nova.scheduler.client.report [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted allocations for instance 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400 [ 903.544437] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.558954] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794797, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.656333] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-b1bd98d9-bd0f-4abd-a188-e5267ada4852-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.656603] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-b1bd98d9-bd0f-4abd-a188-e5267ada4852-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.668326] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794794, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.689642] env[61995]: DEBUG nova.network.neutron [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updated VIF entry in instance network info cache for port 8c280b5d-bf84-4a8d-8ab7-4230645c59d1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.689878] env[61995]: DEBUG nova.network.neutron [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating instance_info_cache with network_info: [{"id": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "address": "fa:16:3e:c6:74:e1", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c280b5d-bf", "ovs_interfaceid": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.785521] env[61995]: DEBUG nova.network.neutron [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Port 9eb488f7-7474-4b91-9b01-2f4378f5daca binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 903.798566] env[61995]: DEBUG nova.network.neutron [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Updated VIF entry in instance network info cache for port 05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.799155] env[61995]: DEBUG nova.network.neutron [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Updating instance_info_cache with network_info: [{"id": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "address": "fa:16:3e:c9:4f:e8", "network": {"id": "3f7f34b5-aa6b-4e83-8593-e51a8f0544e8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-50477412-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "837124ca1f354c8aae242a8af95756a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a69ed1dd-213a-4e30-992a-466735188bf6", "external-id": "nsx-vlan-transportzone-102", "segmentation_id": 102, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05b0b88f-30", "ovs_interfaceid": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.903367] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b1047003-8537-4434-a641-4ce3fda429b4 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "5480efdf-fd46-4a26-b6a5-b6c3dd4c4400" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.588s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.064303] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794797, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.066663] env[61995]: DEBUG nova.network.neutron [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Successfully updated port: a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.167333] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.168295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.171036] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936d8325-b5ed-4163-9a7c-81ac53b6ed7a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.196497] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794794, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.65724} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.197567] env[61995]: DEBUG oslo_concurrency.lockutils [req-710ae382-912a-4b2e-bd86-83d31a34e4c2 req-b8ffd1a3-c24d-45c3-9e51-5bad3858fdb1 service nova] Releasing lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.200711] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] a8b7bcd6-4753-4e1f-8b0b-777c882b99f5/a8b7bcd6-4753-4e1f-8b0b-777c882b99f5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 904.201082] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.201881] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efcf5ea-28c0-46de-a207-7755f82af0f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.206072] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec186418-303e-42bb-ab2b-73c9d9460a41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.214767] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 904.214767] env[61995]: value = "task-794798" [ 904.214767] env[61995]: _type = "Task" [ 904.214767] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.243095] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Reconfiguring VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 904.247296] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2595d4e-2c4e-4cc7-a016-ecfc5719f293 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.270235] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.271792] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 904.271792] env[61995]: value = "task-794799" [ 904.271792] env[61995]: _type = "Task" [ 904.271792] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.276394] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceaa1320-cac0-41dd-93e1-20fd73f91b67 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.289543] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d96793f-b17b-4b1a-b108-1bf810d5dd1f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.293763] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.329495] env[61995]: DEBUG oslo_concurrency.lockutils [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] Releasing lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.329762] env[61995]: DEBUG nova.compute.manager [req-5014f2d9-1a35-48d3-bd4d-8ec58b8f7750 req-718c041e-f023-4a39-809f-17b57f8ba14d service nova] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Received event network-vif-deleted-a410e471-f28f-4398-81c2-d4cd2bc69ba8 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.331329] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b418ed0-fb7a-4934-bea0-63066b0e9fa7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.340370] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda6a837-410a-4994-850e-a12abf138c0c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.356078] env[61995]: DEBUG nova.compute.provider_tree [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.548296] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.548668] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.548931] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.549185] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.549413] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.555401] env[61995]: INFO nova.compute.manager [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Terminating instance [ 904.557632] env[61995]: DEBUG nova.compute.manager [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 904.557864] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 904.558726] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb962ffc-3b3c-44a6-987d-c0c316cf34c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.565101] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794797, 'name': CreateVM_Task, 'duration_secs': 1.351185} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.565652] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 904.566368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.566563] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.566912] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.567200] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c06f9747-b792-4d8c-9e14-92dd77fcc08d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.571131] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Acquiring lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.571339] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Acquired lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.571466] env[61995]: DEBUG nova.network.neutron [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 904.572641] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 904.573416] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eabb102b-e066-41fd-bc40-83c3f1967f0d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.577028] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 904.577028] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52fc5539-c3a6-37e7-b4f5-820ecc09dee7" [ 904.577028] env[61995]: _type = "Task" [ 904.577028] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.582572] env[61995]: DEBUG oslo_vmware.api [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 904.582572] env[61995]: value = "task-794800" [ 904.582572] env[61995]: _type = "Task" [ 904.582572] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.590309] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fc5539-c3a6-37e7-b4f5-820ecc09dee7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.597064] env[61995]: DEBUG oslo_vmware.api [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.747137] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122834} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.747433] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.748251] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f535040-2be9-4efb-9d40-6bbc21293105 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.774782] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] a8b7bcd6-4753-4e1f-8b0b-777c882b99f5/a8b7bcd6-4753-4e1f-8b0b-777c882b99f5.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.775206] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13223401-d56f-47de-bf2a-bf4bff559a9e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.809860] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.815011] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 904.815011] env[61995]: value = "task-794801" [ 904.815011] env[61995]: _type = "Task" [ 904.815011] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.819449] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.819746] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.820065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.832460] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.860302] env[61995]: DEBUG nova.scheduler.client.report [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.870566] env[61995]: DEBUG nova.compute.manager [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Received event network-vif-plugged-a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.870752] env[61995]: DEBUG oslo_concurrency.lockutils [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] Acquiring lock "7d987006-afac-4973-ae69-f15b4752469b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.871924] env[61995]: DEBUG oslo_concurrency.lockutils [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] Lock "7d987006-afac-4973-ae69-f15b4752469b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.871924] env[61995]: DEBUG oslo_concurrency.lockutils [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] Lock "7d987006-afac-4973-ae69-f15b4752469b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.871924] env[61995]: DEBUG nova.compute.manager [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] No waiting events found dispatching network-vif-plugged-a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.871924] env[61995]: WARNING nova.compute.manager [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Received unexpected event network-vif-plugged-a29a835c-8bc8-4c90-b691-44a040b5c90f for instance with vm_state building and task_state spawning. [ 904.871924] env[61995]: DEBUG nova.compute.manager [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Received event network-changed-a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.872129] env[61995]: DEBUG nova.compute.manager [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Refreshing instance network info cache due to event network-changed-a29a835c-8bc8-4c90-b691-44a040b5c90f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.872262] env[61995]: DEBUG oslo_concurrency.lockutils [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] Acquiring lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.092050] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fc5539-c3a6-37e7-b4f5-820ecc09dee7, 'name': SearchDatastore_Task, 'duration_secs': 0.01954} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.092840] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.093242] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.093477] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.093924] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.093924] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.094198] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e027f3e-9cd3-404c-8c6f-631b246452a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.099567] env[61995]: DEBUG oslo_vmware.api [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794800, 'name': PowerOffVM_Task, 'duration_secs': 0.447018} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.100327] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 905.100493] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 905.100755] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65594517-c6d4-48c4-a156-69b0fbf9f575 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.108513] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.108906] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 905.110800] env[61995]: DEBUG nova.network.neutron [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 905.113471] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb9295a-1d0c-43f3-965f-fafe50836a18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.122811] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 905.122811] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c7c283-748a-7115-9359-81faa7ab26c4" [ 905.122811] env[61995]: _type = "Task" [ 905.122811] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.131223] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c7c283-748a-7115-9359-81faa7ab26c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.146695] env[61995]: DEBUG nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.147252] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.147470] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.147629] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.147812] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.147966] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.148207] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.148426] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.148589] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.148778] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.148951] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.149223] env[61995]: DEBUG nova.virt.hardware [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.150080] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534a2450-b7a2-40e0-83bc-1a4f4120773e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.161711] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a7e8ce-b6a9-4690-83ce-662fc2436831 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.197585] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 905.197829] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 905.198162] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleting the datastore file [datastore1] 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.198397] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8309c658-1c37-4fac-b446-d9d9cb065815 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.208284] env[61995]: DEBUG oslo_vmware.api [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 905.208284] env[61995]: value = "task-794803" [ 905.208284] env[61995]: _type = "Task" [ 905.208284] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.219490] env[61995]: DEBUG oslo_vmware.api [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.286808] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.334882] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794801, 'name': ReconfigVM_Task, 'duration_secs': 0.526092} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.334882] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Reconfigured VM instance instance-00000048 to attach disk [datastore1] a8b7bcd6-4753-4e1f-8b0b-777c882b99f5/a8b7bcd6-4753-4e1f-8b0b-777c882b99f5.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.335330] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-969b257d-87df-458e-ab56-1f1acf2c46cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.343915] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 905.343915] env[61995]: value = "task-794804" [ 905.343915] env[61995]: _type = "Task" [ 905.343915] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.356782] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794804, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.363147] env[61995]: DEBUG nova.network.neutron [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updating instance_info_cache with network_info: [{"id": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "address": "fa:16:3e:78:50:28", "network": {"id": "d92807ce-4e1a-4ce2-92f2-4d091cdce362", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-102839657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ed911ecc2748cc9ec398029d3b3e78", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa29a835c-8b", "ovs_interfaceid": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.634642] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c7c283-748a-7115-9359-81faa7ab26c4, 'name': SearchDatastore_Task, 'duration_secs': 0.015115} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.635515] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3326702-3516-43f8-90f4-9cc00273424f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.642198] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 905.642198] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d85ce1-30c9-307b-7bb7-31b3fbff9048" [ 905.642198] env[61995]: _type = "Task" [ 905.642198] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.651193] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d85ce1-30c9-307b-7bb7-31b3fbff9048, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.720710] env[61995]: DEBUG oslo_vmware.api [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281085} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.721027] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.721232] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 905.721460] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 905.721730] env[61995]: INFO nova.compute.manager [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 905.722103] env[61995]: DEBUG oslo.service.loopingcall [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.722387] env[61995]: DEBUG nova.compute.manager [-] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.722535] env[61995]: DEBUG nova.network.neutron [-] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 905.786990] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.855215] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794804, 'name': Rename_Task, 'duration_secs': 0.223518} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.855500] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 905.855757] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a85616b1-e3b2-4033-b0cc-a93107f006cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.868261] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 905.868261] env[61995]: value = "task-794805" [ 905.868261] env[61995]: _type = "Task" [ 905.868261] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.871527] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.507s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.874325] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Releasing lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.874599] env[61995]: DEBUG nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Instance network_info: |[{"id": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "address": "fa:16:3e:78:50:28", "network": {"id": "d92807ce-4e1a-4ce2-92f2-4d091cdce362", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-102839657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ed911ecc2748cc9ec398029d3b3e78", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa29a835c-8b", "ovs_interfaceid": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.878177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.894s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.878378] env[61995]: DEBUG nova.objects.instance [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 905.880706] env[61995]: DEBUG oslo_concurrency.lockutils [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] Acquired lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.884102] env[61995]: DEBUG nova.network.neutron [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Refreshing network info cache for port a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.884102] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:50:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94926d5b-bfab-4c04-85b5-0fe89934c8ff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a29a835c-8bc8-4c90-b691-44a040b5c90f', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.889492] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Creating folder: Project (d1ed911ecc2748cc9ec398029d3b3e78). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.889784] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.890758] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.890913] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.891094] env[61995]: DEBUG nova.network.neutron [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 905.892490] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b25ddc53-4972-4b88-ba31-eb74b3379347 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.909529] env[61995]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 905.910125] env[61995]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61995) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 905.910424] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Folder already exists: Project (d1ed911ecc2748cc9ec398029d3b3e78). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.910829] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Creating folder: Instances. Parent ref: group-v185336. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.911220] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2e80af7c-cb8b-4ed2-8781-e0c7a97af47b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.929224] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Created folder: Instances in parent group-v185336. [ 905.929502] env[61995]: DEBUG oslo.service.loopingcall [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.929703] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 905.929924] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cef002e6-25e2-4900-9f65-f5ca5029a31c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.952154] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.952154] env[61995]: value = "task-794808" [ 905.952154] env[61995]: _type = "Task" [ 905.952154] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.966274] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794808, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.096883] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6e39a-3c4d-4284-bd0a-59d993d2716e/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 906.097828] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffd49bc-2f6a-4bd1-a38c-6d04fb6c1051 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.105234] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6e39a-3c4d-4284-bd0a-59d993d2716e/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 906.105450] env[61995]: ERROR oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6e39a-3c4d-4284-bd0a-59d993d2716e/disk-0.vmdk due to incomplete transfer. [ 906.105713] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-897b3dd0-303a-49ec-a6bb-d54f4c954858 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.115786] env[61995]: DEBUG oslo_vmware.rw_handles [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6e39a-3c4d-4284-bd0a-59d993d2716e/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 906.115985] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Uploaded image d9aa4058-3449-47b3-b10b-42d75d2fcc13 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 906.119167] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 906.119528] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-50009fd7-51f2-4e9c-9048-03b1b1c4223c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.128067] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 906.128067] env[61995]: value = "task-794809" [ 906.128067] env[61995]: _type = "Task" [ 906.128067] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.138783] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794809, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.153907] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d85ce1-30c9-307b-7bb7-31b3fbff9048, 'name': SearchDatastore_Task, 'duration_secs': 0.01407} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.154203] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.154451] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 5a6bd8a7-2633-42d2-b853-4e3a5690c77e/5a6bd8a7-2633-42d2-b853-4e3a5690c77e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 906.154716] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32823700-0d04-4ec7-97e9-cc87461bb3af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.162481] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 906.162481] env[61995]: value = "task-794810" [ 906.162481] env[61995]: _type = "Task" [ 906.162481] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.171896] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794810, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.287285] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.374987] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794805, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.448321] env[61995]: INFO nova.scheduler.client.report [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted allocation for migration a8348ead-3c45-4010-98f6-841d2a6dec81 [ 906.471414] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794808, 'name': CreateVM_Task, 'duration_secs': 0.416768} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.471414] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 906.471864] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185340', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'name': 'volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d987006-afac-4973-ae69-f15b4752469b', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'serial': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02'}, 'attachment_id': '61bf43d0-0bd1-4d9d-80e9-33372f9da32c', 'guest_format': None, 'boot_index': 0, 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61995) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 906.472143] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Root volume attach. Driver type: vmdk {{(pid=61995) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 906.472979] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54cd05d-752f-45dc-97af-77817834f8ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.482359] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1430a5-ee41-4ac4-a7ae-023fcd036a39 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.490665] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b7a3b6-62e3-42b9-9a86-f6c2c16cd2a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.498814] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-9235d1cd-dac3-4dbd-bb47-ba65697c360e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.525474] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 906.525474] env[61995]: value = "task-794811" [ 906.525474] env[61995]: _type = "Task" [ 906.525474] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.536529] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.564981] env[61995]: DEBUG nova.network.neutron [-] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.639233] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794809, 'name': Destroy_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.677716] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794810, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.790757] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.875638] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794805, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.900096] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1844085f-b8a8-4d52-bdb4-85dd11360fe9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.901300] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.357s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.901534] env[61995]: DEBUG nova.objects.instance [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'resources' on Instance uuid fdeceddf-c6f9-4a98-afd0-defa1284eeff {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.941544] env[61995]: DEBUG nova.network.neutron [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updated VIF entry in instance network info cache for port a29a835c-8bc8-4c90-b691-44a040b5c90f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.941921] env[61995]: DEBUG nova.network.neutron [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updating instance_info_cache with network_info: [{"id": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "address": "fa:16:3e:78:50:28", "network": {"id": "d92807ce-4e1a-4ce2-92f2-4d091cdce362", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-102839657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ed911ecc2748cc9ec398029d3b3e78", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa29a835c-8b", "ovs_interfaceid": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.952483] env[61995]: DEBUG nova.network.neutron [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance_info_cache with network_info: [{"id": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "address": "fa:16:3e:45:26:f9", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb488f7-74", "ovs_interfaceid": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.954921] env[61995]: DEBUG oslo_concurrency.lockutils [None req-51f1fc45-8419-4075-9c73-1c9e2c5e014a tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.177s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.974047] env[61995]: DEBUG nova.compute.manager [req-07488e25-4fef-4dba-ad21-1018a73d3aea req-60727996-11ac-4e31-8833-e644e4ae278c service nova] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Received event network-vif-deleted-9ba91a20-0f20-4a6a-86c4-bbb63bf131fc {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.039283] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 34%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.067830] env[61995]: INFO nova.compute.manager [-] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Took 1.35 seconds to deallocate network for instance. [ 907.148026] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794809, 'name': Destroy_Task, 'duration_secs': 0.650936} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.148672] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Destroyed the VM [ 907.148672] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 907.148972] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a675dcca-d491-454b-87c6-7a3ecc691223 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.158255] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 907.158255] env[61995]: value = "task-794812" [ 907.158255] env[61995]: _type = "Task" [ 907.158255] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.167780] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794812, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.177690] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794810, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613863} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.177977] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 5a6bd8a7-2633-42d2-b853-4e3a5690c77e/5a6bd8a7-2633-42d2-b853-4e3a5690c77e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.178263] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.178540] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-77e0e96f-cfd6-4110-a7c6-54cef8ef2694 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.187537] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 907.187537] env[61995]: value = "task-794813" [ 907.187537] env[61995]: _type = "Task" [ 907.187537] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.200445] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794813, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.291585] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.376675] env[61995]: DEBUG oslo_vmware.api [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794805, 'name': PowerOnVM_Task, 'duration_secs': 1.021413} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.377063] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 907.377306] env[61995]: INFO nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Took 13.36 seconds to spawn the instance on the hypervisor. [ 907.377484] env[61995]: DEBUG nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.378373] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa8bd9e-0c99-4fdf-bd50-4fe5f6224460 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.444381] env[61995]: DEBUG oslo_concurrency.lockutils [req-af549d3e-ac0d-4f51-880a-848104e31e81 req-8d6dc017-7729-4dd5-bc29-c2e37f857376 service nova] Releasing lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.455035] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.540072] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 47%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.575368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.675760] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794812, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.702511] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794813, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13733} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.702788] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.704059] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2a58a6-b25a-4687-999d-19ff57407277 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.734742] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 5a6bd8a7-2633-42d2-b853-4e3a5690c77e/5a6bd8a7-2633-42d2-b853-4e3a5690c77e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.738802] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc77fd97-3c14-4dbc-89c2-8dfe94f76df1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.765491] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 907.765491] env[61995]: value = "task-794814" [ 907.765491] env[61995]: _type = "Task" [ 907.765491] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.775728] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794814, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.790356] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.821835] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf519b2-cc91-4be9-b2e6-9247801abc4e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.831477] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edd5df8-7cc7-43e5-8c0d-dcceb3a233a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.870489] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a8fab5-6a97-468f-9239-bccf2c618925 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.880170] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f45ab7-f5f3-4017-8e4f-32aa1c97c298 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.901794] env[61995]: DEBUG nova.compute.provider_tree [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.907164] env[61995]: INFO nova.compute.manager [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Took 28.86 seconds to build instance. [ 907.989421] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e500cff-9df1-4772-8ca0-17d504bf335e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.010624] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63807852-340f-47ea-aa8d-58d27a732e4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.020213] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance '4b49e7b1-7fd0-4e59-af79-5c0898967c35' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 908.039434] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 60%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.170936] env[61995]: DEBUG nova.compute.manager [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.171417] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794812, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.172849] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd7f15e-b076-4287-b71a-040b153112d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.279098] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794814, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.289926] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.408376] env[61995]: DEBUG nova.scheduler.client.report [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.412579] env[61995]: DEBUG oslo_concurrency.lockutils [None req-69f9f889-73f5-404c-96e0-70963de1c818 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.382s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.526756] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.527114] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f4b0a30-ce29-4f96-af7d-b07fc959e872 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.542628] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 73%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.544268] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 908.544268] env[61995]: value = "task-794815" [ 908.544268] env[61995]: _type = "Task" [ 908.544268] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.553960] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794815, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.671382] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794812, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.684587] env[61995]: INFO nova.compute.manager [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] instance snapshotting [ 908.685378] env[61995]: DEBUG nova.objects.instance [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'flavor' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.778551] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794814, 'name': ReconfigVM_Task, 'duration_secs': 0.698889} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.778868] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 5a6bd8a7-2633-42d2-b853-4e3a5690c77e/5a6bd8a7-2633-42d2-b853-4e3a5690c77e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.779619] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cf4e008-ae71-4120-a60c-996f8edeeeb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.794156] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 908.794156] env[61995]: value = "task-794816" [ 908.794156] env[61995]: _type = "Task" [ 908.794156] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.794437] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.805410] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794816, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.807177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "2b40609a-0826-4a16-af28-c8926af6b646" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.807392] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.807646] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "2b40609a-0826-4a16-af28-c8926af6b646-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.807848] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.808073] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.810430] env[61995]: INFO nova.compute.manager [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Terminating instance [ 908.812335] env[61995]: DEBUG nova.compute.manager [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.812568] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 908.813422] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3dea11-3039-48db-931a-5c161968bbbb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.825511] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 908.825852] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1454d576-adb0-4719-9b04-7f38b75ad2f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.837422] env[61995]: DEBUG oslo_vmware.api [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 908.837422] env[61995]: value = "task-794817" [ 908.837422] env[61995]: _type = "Task" [ 908.837422] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.850268] env[61995]: DEBUG oslo_vmware.api [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.891019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.891019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.891019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.891019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.891019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.892079] env[61995]: INFO nova.compute.manager [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Terminating instance [ 908.894302] env[61995]: DEBUG nova.compute.manager [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.894476] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 908.895480] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faadc22d-9c47-47a7-94d4-a5606c0ea084 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.906959] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 908.907154] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33725fa8-fbdf-4fca-9e7f-575261fc4189 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.914434] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.013s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.919181] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.344s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.919319] env[61995]: DEBUG nova.objects.instance [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lazy-loading 'resources' on Instance uuid 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.920857] env[61995]: DEBUG oslo_vmware.api [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 908.920857] env[61995]: value = "task-794818" [ 908.920857] env[61995]: _type = "Task" [ 908.920857] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.933157] env[61995]: DEBUG oslo_vmware.api [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.942240] env[61995]: INFO nova.scheduler.client.report [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted allocations for instance fdeceddf-c6f9-4a98-afd0-defa1284eeff [ 909.043243] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 86%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.056851] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794815, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.176026] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794812, 'name': RemoveSnapshot_Task, 'duration_secs': 1.574174} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.176422] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 909.177036] env[61995]: DEBUG nova.compute.manager [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.177511] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a64639f-455a-4fa9-847c-58c6a40f45c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.192534] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9802825a-b745-458d-aea8-29290de8ee76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.221378] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f0d240-78b3-4c6d-af46-e70f498d1133 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.295383] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.305550] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794816, 'name': Rename_Task, 'duration_secs': 0.236562} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.305864] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 909.306151] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea956d4d-b0ff-4d35-8d80-560dc4f9583f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.313984] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 909.313984] env[61995]: value = "task-794819" [ 909.313984] env[61995]: _type = "Task" [ 909.313984] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.323268] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.349620] env[61995]: DEBUG oslo_vmware.api [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.438464] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.438806] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.444096] env[61995]: DEBUG oslo_vmware.api [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794818, 'name': PowerOffVM_Task, 'duration_secs': 0.219257} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.444396] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 909.444542] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 909.445772] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-350bc0bb-f86b-42df-88ef-5837e1111b35 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.450351] env[61995]: DEBUG oslo_concurrency.lockutils [None req-10cd788a-69a8-433d-a58b-f76d3d5f621f tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "fdeceddf-c6f9-4a98-afd0-defa1284eeff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.430s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.544852] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.558479] env[61995]: DEBUG oslo_vmware.api [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794815, 'name': PowerOnVM_Task, 'duration_secs': 0.735303} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.558831] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 909.559085] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0c08a8e4-0245-4fb2-a4a2-bfbd7ef7a53e tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance '4b49e7b1-7fd0-4e59-af79-5c0898967c35' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 909.695991] env[61995]: INFO nova.compute.manager [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Shelve offloading [ 909.698058] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.699152] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e3b4111-d913-4786-b7c7-2512ff87d10b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.709153] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 909.709153] env[61995]: value = "task-794821" [ 909.709153] env[61995]: _type = "Task" [ 909.709153] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.723411] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 909.723411] env[61995]: DEBUG nova.compute.manager [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.724136] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd495f1d-806c-45bf-8245-e12c9ce8e243 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.732492] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.732713] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.732838] env[61995]: DEBUG nova.network.neutron [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 909.737049] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 909.737440] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e674e2f7-863f-4cec-9658-e4db80f51fef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.745285] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de077ef-f0e2-4604-b650-378521e3a302 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.749222] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 909.749222] env[61995]: value = "task-794822" [ 909.749222] env[61995]: _type = "Task" [ 909.749222] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.755938] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c508dae-5898-460a-856a-ce3210ee30a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.761968] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794822, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.793094] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3e8901-9ae5-4c4a-84f8-d32b2fd2b853 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.801767] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.805158] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973be71a-523b-4e6a-bda5-fbb4229ad8be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.820184] env[61995]: DEBUG nova.compute.provider_tree [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.830851] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794819, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.848340] env[61995]: DEBUG oslo_vmware.api [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794817, 'name': PowerOffVM_Task, 'duration_secs': 0.68903} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.848627] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 909.848833] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 909.849076] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b161270-2983-4c60-953a-ec97d68c4b82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.936372] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 909.936629] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 909.936821] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleting the datastore file [datastore2] 2b40609a-0826-4a16-af28-c8926af6b646 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.937152] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11aa87a8-7394-46db-b10a-fa0077b2c043 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.941127] env[61995]: DEBUG nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 909.945253] env[61995]: DEBUG oslo_vmware.api [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 909.945253] env[61995]: value = "task-794824" [ 909.945253] env[61995]: _type = "Task" [ 909.945253] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.955306] env[61995]: DEBUG oslo_vmware.api [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.042410] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.259449] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794822, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.301167] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.325422] env[61995]: DEBUG nova.scheduler.client.report [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.335624] env[61995]: DEBUG oslo_vmware.api [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794819, 'name': PowerOnVM_Task, 'duration_secs': 0.57347} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.336069] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.336215] env[61995]: INFO nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Took 10.51 seconds to spawn the instance on the hypervisor. [ 910.336321] env[61995]: DEBUG nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.337142] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2754a9e-87e8-4ff1-a856-e39f67c344cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.461590] env[61995]: DEBUG oslo_vmware.api [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.48765} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.462090] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.462323] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 910.462521] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 910.463155] env[61995]: INFO nova.compute.manager [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Took 1.65 seconds to destroy the instance on the hypervisor. [ 910.463155] env[61995]: DEBUG oslo.service.loopingcall [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.463384] env[61995]: DEBUG nova.compute.manager [-] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.463496] env[61995]: DEBUG nova.network.neutron [-] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 910.470614] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.531097] env[61995]: DEBUG nova.network.neutron [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee71b78-02", "ovs_interfaceid": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.542960] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 98%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.760756] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794822, 'name': CreateSnapshot_Task, 'duration_secs': 0.621293} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.761071] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 910.761828] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fbdc6d-2ad4-4c4d-97cf-b617d08a9c3f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.801048] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.832051] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.834553] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.364s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.839383] env[61995]: INFO nova.compute.claims [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.854524] env[61995]: INFO nova.scheduler.client.report [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted allocations for instance 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d [ 910.866019] env[61995]: INFO nova.compute.manager [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Took 27.38 seconds to build instance. [ 910.881145] env[61995]: DEBUG nova.compute.manager [req-2edaded5-39b1-4585-991f-22688f239a8d req-65efeb75-4a84-4086-b478-94a1c9e2c8e4 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Received event network-vif-deleted-3ecd14b2-d6b9-4216-936e-edae029037af {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.881145] env[61995]: INFO nova.compute.manager [req-2edaded5-39b1-4585-991f-22688f239a8d req-65efeb75-4a84-4086-b478-94a1c9e2c8e4 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Neutron deleted interface 3ecd14b2-d6b9-4216-936e-edae029037af; detaching it from the instance and deleting it from the info cache [ 910.881145] env[61995]: DEBUG nova.network.neutron [req-2edaded5-39b1-4585-991f-22688f239a8d req-65efeb75-4a84-4086-b478-94a1c9e2c8e4 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.037034] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.042809] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 98%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.279668] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 911.280082] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ec719206-c7a8-4206-91b8-794832fc5504 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.292596] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 911.292596] env[61995]: value = "task-794825" [ 911.292596] env[61995]: _type = "Task" [ 911.292596] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.306577] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794825, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.308760] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.328335] env[61995]: DEBUG nova.network.neutron [-] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.370613] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f17cc542-0631-4a0c-96b4-7437ea839e8d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "0c45e1fc-7b9a-4873-a381-ee2850ad5d4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.822s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.371844] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14f75dc4-7931-46ff-b144-e5a38680d19e tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.895s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.383308] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aed5e030-da8e-4e2e-942c-c9c542976e41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.395616] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6383fe28-319f-43ef-ab16-f85f6e3b0d53 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.417301] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 911.418346] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abda44a-5008-4768-8619-452369731a6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.437331] env[61995]: DEBUG nova.compute.manager [req-2edaded5-39b1-4585-991f-22688f239a8d req-65efeb75-4a84-4086-b478-94a1c9e2c8e4 service nova] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Detach interface failed, port_id=3ecd14b2-d6b9-4216-936e-edae029037af, reason: Instance 2b40609a-0826-4a16-af28-c8926af6b646 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 911.440135] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 911.442294] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d80ea80-9b92-46af-bf34-09d3be4469e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.546765] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task} progress is 98%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.617305] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 911.617549] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 911.617746] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleting the datastore file [datastore1] a8b7bcd6-4753-4e1f-8b0b-777c882b99f5 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.618073] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a2eeab9-5855-4fb8-8ed8-fcf7862f1197 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.626702] env[61995]: DEBUG oslo_vmware.api [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 911.626702] env[61995]: value = "task-794827" [ 911.626702] env[61995]: _type = "Task" [ 911.626702] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.634838] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 911.635098] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 911.635315] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleting the datastore file [datastore1] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.638939] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7399023-756c-40d2-9974-9eceb7e12cdf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.641036] env[61995]: DEBUG oslo_vmware.api [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.647064] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 911.647064] env[61995]: value = "task-794828" [ 911.647064] env[61995]: _type = "Task" [ 911.647064] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.659292] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.713556] env[61995]: DEBUG nova.network.neutron [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Port 9eb488f7-7474-4b91-9b01-2f4378f5daca binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 911.714789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.714789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.714789] env[61995]: DEBUG nova.network.neutron [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.814826] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794825, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.821021] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.831828] env[61995]: INFO nova.compute.manager [-] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Took 1.37 seconds to deallocate network for instance. [ 911.952876] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.952876] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.952876] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.952876] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.952876] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.959078] env[61995]: INFO nova.compute.manager [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Terminating instance [ 911.965258] env[61995]: DEBUG nova.compute.manager [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 911.965571] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 911.966052] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.966404] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.967807] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1b6918-f25b-4151-a117-7d5026a23e81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.980514] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 911.980514] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52ef32e7-7657-471c-bc99-d017fbbcda21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.989030] env[61995]: DEBUG oslo_vmware.api [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 911.989030] env[61995]: value = "task-794829" [ 911.989030] env[61995]: _type = "Task" [ 911.989030] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.997912] env[61995]: DEBUG oslo_vmware.api [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794829, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.045062] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794811, 'name': RelocateVM_Task, 'duration_secs': 5.321686} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.047711] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 912.047932] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185340', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'name': 'volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d987006-afac-4973-ae69-f15b4752469b', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'serial': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 912.049093] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef29475-e4a7-462d-b788-8a1cd6502192 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.068436] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d258e9-bd8a-47ae-b0b4-4b3af6729b0f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.092938] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02/volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.095815] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f515f287-39ba-4f8f-afc3-a293c3c4d732 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.118770] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 912.118770] env[61995]: value = "task-794830" [ 912.118770] env[61995]: _type = "Task" [ 912.118770] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.131486] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794830, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.143785] env[61995]: DEBUG oslo_vmware.api [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794827, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178383} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.144222] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.144366] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 912.144483] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 912.144684] env[61995]: INFO nova.compute.manager [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Took 3.25 seconds to destroy the instance on the hypervisor. [ 912.144972] env[61995]: DEBUG oslo.service.loopingcall [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.145168] env[61995]: DEBUG nova.compute.manager [-] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 912.145268] env[61995]: DEBUG nova.network.neutron [-] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 912.160434] env[61995]: DEBUG oslo_vmware.api [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794828, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182408} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.163659] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.163907] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 912.164218] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 912.189034] env[61995]: INFO nova.scheduler.client.report [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleted allocations for instance 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 [ 912.255098] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe92543-a2be-427d-8f16-f68f77874659 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.263860] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f6eb27-9d63-4eae-bdf5-79aea011f53b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.306181] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c26734-40de-44eb-8c8f-67f90d25ab66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.315046] env[61995]: DEBUG oslo_vmware.api [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794799, 'name': ReconfigVM_Task, 'duration_secs': 7.642635} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.322444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.322698] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Reconfigured VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 912.325242] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794825, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.326611] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e3b1bb-5578-433f-86d4-205aba3b1080 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.342648] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.343213] env[61995]: DEBUG nova.compute.provider_tree [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.474311] env[61995]: DEBUG nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.501354] env[61995]: DEBUG oslo_vmware.api [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794829, 'name': PowerOffVM_Task, 'duration_secs': 0.290059} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.501645] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 912.501827] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 912.502118] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46f1db21-9812-433e-a352-fc22669007c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.635123] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794830, 'name': ReconfigVM_Task, 'duration_secs': 0.380384} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.641552] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Reconfigured VM instance instance-0000004a to attach disk [datastore1] volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02/volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.641552] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37988760-890e-43cf-b18b-7209f09478ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.653102] env[61995]: DEBUG nova.network.neutron [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance_info_cache with network_info: [{"id": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "address": "fa:16:3e:45:26:f9", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb488f7-74", "ovs_interfaceid": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.662802] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 912.662802] env[61995]: value = "task-794832" [ 912.662802] env[61995]: _type = "Task" [ 912.662802] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.675551] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794832, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.683008] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 912.683311] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 912.683518] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleting the datastore file [datastore1] fcdc7c33-8c46-4f4d-bb39-e6f943256cb0 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.683787] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42d4304c-60d6-49ff-8cd2-394b4b7bac8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.690710] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.692476] env[61995]: DEBUG oslo_vmware.api [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 912.692476] env[61995]: value = "task-794833" [ 912.692476] env[61995]: _type = "Task" [ 912.692476] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.695182] env[61995]: DEBUG nova.compute.manager [req-42334ff1-cf1d-4b8a-a037-2af359d69405 req-5ca11eb7-8e62-40e6-8af0-f956bf3ad538 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received event network-vif-deleted-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.695412] env[61995]: INFO nova.compute.manager [req-42334ff1-cf1d-4b8a-a037-2af359d69405 req-5ca11eb7-8e62-40e6-8af0-f956bf3ad538 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Neutron deleted interface 7812f1a4-43be-4dc0-82a1-b92cd78fb1e1; detaching it from the instance and deleting it from the info cache [ 912.695674] env[61995]: DEBUG nova.network.neutron [req-42334ff1-cf1d-4b8a-a037-2af359d69405 req-5ca11eb7-8e62-40e6-8af0-f956bf3ad538 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.706793] env[61995]: DEBUG oslo_vmware.api [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.809775] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794825, 'name': CloneVM_Task, 'duration_secs': 1.284889} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.810064] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Created linked-clone VM from snapshot [ 912.810863] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab96bed-aab9-4f9f-a53e-7324f358414a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.819130] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Uploading image 4c8c4eb7-8495-4e66-b897-ea812737ebc8 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 912.841229] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 912.841229] env[61995]: value = "vm-185358" [ 912.841229] env[61995]: _type = "VirtualMachine" [ 912.841229] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 912.841825] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ee251070-66d1-45dd-bc14-79a3db82626d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.846651] env[61995]: DEBUG nova.scheduler.client.report [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.854033] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease: (returnval){ [ 912.854033] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a99d63-e93d-c10b-047a-65c849721c1b" [ 912.854033] env[61995]: _type = "HttpNfcLease" [ 912.854033] env[61995]: } obtained for exporting VM: (result){ [ 912.854033] env[61995]: value = "vm-185358" [ 912.854033] env[61995]: _type = "VirtualMachine" [ 912.854033] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 912.854033] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the lease: (returnval){ [ 912.854033] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a99d63-e93d-c10b-047a-65c849721c1b" [ 912.854033] env[61995]: _type = "HttpNfcLease" [ 912.854033] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 912.862594] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 912.862594] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a99d63-e93d-c10b-047a-65c849721c1b" [ 912.862594] env[61995]: _type = "HttpNfcLease" [ 912.862594] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 912.911357] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received event network-vif-unplugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.911357] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.911357] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.911357] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.911649] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] No waiting events found dispatching network-vif-unplugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.911746] env[61995]: WARNING nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received unexpected event network-vif-unplugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c for instance with vm_state shelved_offloaded and task_state None. [ 912.911912] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received event network-changed-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.912069] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Refreshing instance network info cache due to event network-changed-6ee71b78-02c5-4f63-a764-d5f1bb61107c. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 912.912324] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Acquiring lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.912450] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Acquired lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.912614] env[61995]: DEBUG nova.network.neutron [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Refreshing network info cache for port 6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.996818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.126589] env[61995]: DEBUG nova.network.neutron [-] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.155506] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.173786] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794832, 'name': ReconfigVM_Task, 'duration_secs': 0.283239} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.174388] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185340', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'name': 'volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d987006-afac-4973-ae69-f15b4752469b', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'serial': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 913.174977] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a38560f-1a49-4ef7-9160-0b4c44724b0f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.184888] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 913.184888] env[61995]: value = "task-794835" [ 913.184888] env[61995]: _type = "Task" [ 913.184888] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.200753] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794835, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.201753] env[61995]: DEBUG oslo_concurrency.lockutils [req-42334ff1-cf1d-4b8a-a037-2af359d69405 req-5ca11eb7-8e62-40e6-8af0-f956bf3ad538 service nova] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.201929] env[61995]: DEBUG oslo_concurrency.lockutils [req-42334ff1-cf1d-4b8a-a037-2af359d69405 req-5ca11eb7-8e62-40e6-8af0-f956bf3ad538 service nova] Acquired lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.202688] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839bb5b4-a060-40c0-be4a-a640193bc784 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.208358] env[61995]: DEBUG oslo_vmware.api [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.45362} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.209110] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.209391] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 913.209596] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 913.209832] env[61995]: INFO nova.compute.manager [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Took 1.24 seconds to destroy the instance on the hypervisor. [ 913.210207] env[61995]: DEBUG oslo.service.loopingcall [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.210482] env[61995]: DEBUG nova.compute.manager [-] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 913.210617] env[61995]: DEBUG nova.network.neutron [-] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 913.224912] env[61995]: DEBUG oslo_concurrency.lockutils [req-42334ff1-cf1d-4b8a-a037-2af359d69405 req-5ca11eb7-8e62-40e6-8af0-f956bf3ad538 service nova] Releasing lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.225230] env[61995]: WARNING nova.compute.manager [req-42334ff1-cf1d-4b8a-a037-2af359d69405 req-5ca11eb7-8e62-40e6-8af0-f956bf3ad538 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Detach interface failed, port_id=7812f1a4-43be-4dc0-82a1-b92cd78fb1e1, reason: No device with interface-id 7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 exists on VM: nova.exception.NotFound: No device with interface-id 7812f1a4-43be-4dc0-82a1-b92cd78fb1e1 exists on VM [ 913.352728] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.353358] env[61995]: DEBUG nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.356441] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.014s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.356684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.358860] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.668s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.359179] env[61995]: DEBUG nova.objects.instance [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lazy-loading 'resources' on Instance uuid 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.368962] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.368962] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a99d63-e93d-c10b-047a-65c849721c1b" [ 913.368962] env[61995]: _type = "HttpNfcLease" [ 913.368962] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 913.369283] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 913.369283] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a99d63-e93d-c10b-047a-65c849721c1b" [ 913.369283] env[61995]: _type = "HttpNfcLease" [ 913.369283] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 913.370062] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664db604-d3ea-479e-94f7-7d3266c8d9c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.383794] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525b8c81-49f3-b5e2-be11-e81c3bde3e58/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 913.383893] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525b8c81-49f3-b5e2-be11-e81c3bde3e58/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 913.449199] env[61995]: INFO nova.scheduler.client.report [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted allocations for instance 2b40609a-0826-4a16-af28-c8926af6b646 [ 913.531100] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ecedc5de-a7fc-42e4-b19d-b4b1ad585b57 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.631601] env[61995]: INFO nova.compute.manager [-] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Took 1.48 seconds to deallocate network for instance. [ 913.662497] env[61995]: DEBUG nova.compute.manager [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61995) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 913.662768] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.674614] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.674794] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.674965] env[61995]: DEBUG nova.network.neutron [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.695177] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794835, 'name': Rename_Task, 'duration_secs': 0.260873} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.696533] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.701022] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e364947f-5470-443a-b0b8-a1702722ec66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.707503] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 913.707503] env[61995]: value = "task-794836" [ 913.707503] env[61995]: _type = "Task" [ 913.707503] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.716330] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794836, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.793749] env[61995]: DEBUG nova.network.neutron [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updated VIF entry in instance network info cache for port 6ee71b78-02c5-4f63-a764-d5f1bb61107c. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.794162] env[61995]: DEBUG nova.network.neutron [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": null, "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap6ee71b78-02", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.865148] env[61995]: DEBUG nova.compute.utils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.866925] env[61995]: DEBUG nova.objects.instance [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lazy-loading 'numa_topology' on Instance uuid 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.868631] env[61995]: DEBUG nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.868691] env[61995]: DEBUG nova.network.neutron [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 913.913286] env[61995]: DEBUG nova.policy [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2388a3bd1587413285ff9f9ec07739b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9301d14211a1464eb740d9bb745b631a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 913.959271] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1672a8ba-fada-4cef-b109-e393baf86016 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "2b40609a-0826-4a16-af28-c8926af6b646" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.152s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.994596] env[61995]: DEBUG nova.network.neutron [-] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.003745] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.004066] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.004461] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.004666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.004848] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.007125] env[61995]: INFO nova.compute.manager [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Terminating instance [ 914.009322] env[61995]: DEBUG nova.compute.manager [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.009575] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.011162] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12aa37db-0821-45e6-80d3-1ad559f54b92 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.020405] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.020740] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0330d0a4-7f02-413b-b0b7-540c5927c33a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.029710] env[61995]: DEBUG oslo_vmware.api [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 914.029710] env[61995]: value = "task-794837" [ 914.029710] env[61995]: _type = "Task" [ 914.029710] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.039635] env[61995]: DEBUG oslo_vmware.api [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794837, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.137454] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.222789] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794836, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.275634] env[61995]: DEBUG nova.network.neutron [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Successfully created port: 0496010d-b03c-4002-b2d4-372cdbededa9 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.297381] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Releasing lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.297623] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Received event network-changed-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.297820] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Refreshing instance network info cache due to event network-changed-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 914.298725] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Acquiring lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.298995] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Acquired lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.299196] env[61995]: DEBUG nova.network.neutron [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Refreshing network info cache for port 05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 914.372573] env[61995]: DEBUG nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.377489] env[61995]: DEBUG nova.objects.base [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Object Instance<0e152049-f9ca-4ac1-b12d-ae1ad4623ff5> lazy-loaded attributes: resources,numa_topology {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 914.500636] env[61995]: INFO nova.compute.manager [-] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Took 1.29 seconds to deallocate network for instance. [ 914.534725] env[61995]: DEBUG nova.network.neutron [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [{"id": "0f5a4362-281d-470a-95b7-6214553bfc8a", "address": "fa:16:3e:9e:ac:a7", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f5a4362-28", "ovs_interfaceid": "0f5a4362-281d-470a-95b7-6214553bfc8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.544641] env[61995]: DEBUG oslo_vmware.api [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794837, 'name': PowerOffVM_Task, 'duration_secs': 0.263879} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.544641] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 914.544641] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 914.544880] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba1163e8-3e17-4975-8431-7f8b021e0fc9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.634236] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.641763] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 914.645041] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 914.645041] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleting the datastore file [datastore1] b1bd98d9-bd0f-4abd-a188-e5267ada4852 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.645041] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63e921ce-c39c-4881-98cb-6bb21b702a48 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.652082] env[61995]: DEBUG oslo_vmware.api [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 914.652082] env[61995]: value = "task-794839" [ 914.652082] env[61995]: _type = "Task" [ 914.652082] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.660981] env[61995]: DEBUG oslo_vmware.api [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.717634] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef58fb5-9cac-4b64-a11d-f05bc20aaf30 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.726216] env[61995]: DEBUG nova.compute.manager [req-e725f2a1-cf83-4c8d-b235-7756047927be req-5d64ad3c-91ec-4b03-ab56-3789e552e91e service nova] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Received event network-vif-deleted-654127d3-17e0-4939-b4b5-6d2e905f4f35 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.726975] env[61995]: DEBUG oslo_vmware.api [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794836, 'name': PowerOnVM_Task, 'duration_secs': 0.625482} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.729060] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 914.729060] env[61995]: INFO nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Took 9.58 seconds to spawn the instance on the hypervisor. [ 914.729060] env[61995]: DEBUG nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.730046] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3def701-490b-416a-8159-a9e6b4aab6ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.736490] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506553e9-45c0-452e-af7f-b2b4e3736b59 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.773856] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfad1728-2e6e-42eb-9b3a-1c78e883e356 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.784252] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7ff475-f73d-40de-8b95-66e35c8ea0f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.800977] env[61995]: DEBUG nova.compute.provider_tree [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.010440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.038752] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-b1bd98d9-bd0f-4abd-a188-e5267ada4852" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.065562] env[61995]: DEBUG nova.network.neutron [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Updated VIF entry in instance network info cache for port 05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 915.066273] env[61995]: DEBUG nova.network.neutron [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Updating instance_info_cache with network_info: [{"id": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "address": "fa:16:3e:c9:4f:e8", "network": {"id": "3f7f34b5-aa6b-4e83-8593-e51a8f0544e8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-50477412-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "837124ca1f354c8aae242a8af95756a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a69ed1dd-213a-4e30-992a-466735188bf6", "external-id": "nsx-vlan-transportzone-102", "segmentation_id": 102, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05b0b88f-30", "ovs_interfaceid": "05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.162558] env[61995]: DEBUG oslo_vmware.api [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.441436} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.162947] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.163317] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 915.163525] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 915.163865] env[61995]: INFO nova.compute.manager [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Took 1.15 seconds to destroy the instance on the hypervisor. [ 915.164157] env[61995]: DEBUG oslo.service.loopingcall [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.164375] env[61995]: DEBUG nova.compute.manager [-] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.165034] env[61995]: DEBUG nova.network.neutron [-] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.257211] env[61995]: INFO nova.compute.manager [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Took 20.75 seconds to build instance. [ 915.308954] env[61995]: DEBUG nova.scheduler.client.report [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.389281] env[61995]: DEBUG nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 915.420643] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.420643] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.420643] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.420643] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.424086] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.424086] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.424086] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.424086] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.424086] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.424086] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.424086] env[61995]: DEBUG nova.virt.hardware [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.424086] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3997b527-ccbb-4b79-9cad-2262f570b1c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.433591] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748de1dc-ac5f-4b9c-bfd1-70f87fffa794 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.544783] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d7fba482-c714-407b-90ce-f4a6c9eaf2d4 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-b1bd98d9-bd0f-4abd-a188-e5267ada4852-7812f1a4-43be-4dc0-82a1-b92cd78fb1e1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.887s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.570567] env[61995]: DEBUG oslo_concurrency.lockutils [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] Releasing lock "refresh_cache-5a6bd8a7-2633-42d2-b853-4e3a5690c77e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.570567] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Received event network-vif-deleted-14cad421-d68a-4eb8-ad41-f42f7a4d644f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.570567] env[61995]: INFO nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Neutron deleted interface 14cad421-d68a-4eb8-ad41-f42f7a4d644f; detaching it from the instance and deleting it from the info cache [ 915.570567] env[61995]: DEBUG nova.network.neutron [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.747243] env[61995]: DEBUG nova.compute.manager [req-af7b9f21-e72a-4ef6-b972-ad5c712f76a5 req-8ce8aecc-f47e-462e-9b95-ec4b9ad627f7 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Received event network-vif-plugged-0496010d-b03c-4002-b2d4-372cdbededa9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.747243] env[61995]: DEBUG oslo_concurrency.lockutils [req-af7b9f21-e72a-4ef6-b972-ad5c712f76a5 req-8ce8aecc-f47e-462e-9b95-ec4b9ad627f7 service nova] Acquiring lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.747243] env[61995]: DEBUG oslo_concurrency.lockutils [req-af7b9f21-e72a-4ef6-b972-ad5c712f76a5 req-8ce8aecc-f47e-462e-9b95-ec4b9ad627f7 service nova] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.747243] env[61995]: DEBUG oslo_concurrency.lockutils [req-af7b9f21-e72a-4ef6-b972-ad5c712f76a5 req-8ce8aecc-f47e-462e-9b95-ec4b9ad627f7 service nova] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.747243] env[61995]: DEBUG nova.compute.manager [req-af7b9f21-e72a-4ef6-b972-ad5c712f76a5 req-8ce8aecc-f47e-462e-9b95-ec4b9ad627f7 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] No waiting events found dispatching network-vif-plugged-0496010d-b03c-4002-b2d4-372cdbededa9 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.747243] env[61995]: WARNING nova.compute.manager [req-af7b9f21-e72a-4ef6-b972-ad5c712f76a5 req-8ce8aecc-f47e-462e-9b95-ec4b9ad627f7 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Received unexpected event network-vif-plugged-0496010d-b03c-4002-b2d4-372cdbededa9 for instance with vm_state building and task_state spawning. [ 915.759442] env[61995]: DEBUG oslo_concurrency.lockutils [None req-598a99d6-6509-449e-8d75-1807963ca954 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "7d987006-afac-4973-ae69-f15b4752469b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.266s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.813768] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.455s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.821385] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.823s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.821385] env[61995]: INFO nova.compute.claims [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.928065] env[61995]: DEBUG nova.network.neutron [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Successfully updated port: 0496010d-b03c-4002-b2d4-372cdbededa9 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.073114] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc8aac5f-daae-434f-8e67-5c3be456f0e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.085894] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968d5dc2-3c0e-4273-9168-4f9f2c9dad12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.130552] env[61995]: DEBUG nova.compute.manager [req-d56419bf-062b-4847-9163-8cbb9a107e37 req-8b16ab32-5b8b-424b-a788-a470f7d949f0 service nova] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Detach interface failed, port_id=14cad421-d68a-4eb8-ad41-f42f7a4d644f, reason: Instance a8b7bcd6-4753-4e1f-8b0b-777c882b99f5 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 916.323840] env[61995]: DEBUG oslo_concurrency.lockutils [None req-20b6a46d-4e54-4aab-834a-0613df31478b tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.954s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.324969] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.691s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.325244] env[61995]: INFO nova.compute.manager [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Unshelving [ 916.355440] env[61995]: DEBUG nova.network.neutron [-] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.432620] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.432620] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.432620] env[61995]: DEBUG nova.network.neutron [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 916.755143] env[61995]: DEBUG nova.compute.manager [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Received event network-vif-deleted-0f5a4362-281d-470a-95b7-6214553bfc8a {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.755348] env[61995]: DEBUG nova.compute.manager [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Received event network-changed-a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.755523] env[61995]: DEBUG nova.compute.manager [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Refreshing instance network info cache due to event network-changed-a29a835c-8bc8-4c90-b691-44a040b5c90f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 916.755831] env[61995]: DEBUG oslo_concurrency.lockutils [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] Acquiring lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.756047] env[61995]: DEBUG oslo_concurrency.lockutils [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] Acquired lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.756241] env[61995]: DEBUG nova.network.neutron [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Refreshing network info cache for port a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 916.856675] env[61995]: INFO nova.compute.manager [-] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Took 1.69 seconds to deallocate network for instance. [ 916.977726] env[61995]: DEBUG nova.network.neutron [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 917.153249] env[61995]: DEBUG nova.network.neutron [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance_info_cache with network_info: [{"id": "0496010d-b03c-4002-b2d4-372cdbededa9", "address": "fa:16:3e:6d:e6:cd", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0496010d-b0", "ovs_interfaceid": "0496010d-b03c-4002-b2d4-372cdbededa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.166482] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6ce9e6-adda-479c-a49d-c373d75a4087 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.176139] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702b1606-5ddf-4f1b-b218-5a9b41186026 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.210500] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fe1759-54ba-4555-8e10-c78fb378bd6f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.219827] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3250a4d3-e849-44a8-87aa-d183478a6ecd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.234992] env[61995]: DEBUG nova.compute.provider_tree [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.355796] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.368186] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.504171] env[61995]: DEBUG nova.network.neutron [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updated VIF entry in instance network info cache for port a29a835c-8bc8-4c90-b691-44a040b5c90f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.504581] env[61995]: DEBUG nova.network.neutron [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updating instance_info_cache with network_info: [{"id": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "address": "fa:16:3e:78:50:28", "network": {"id": "d92807ce-4e1a-4ce2-92f2-4d091cdce362", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-102839657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ed911ecc2748cc9ec398029d3b3e78", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa29a835c-8b", "ovs_interfaceid": "a29a835c-8bc8-4c90-b691-44a040b5c90f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.656204] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.656561] env[61995]: DEBUG nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Instance network_info: |[{"id": "0496010d-b03c-4002-b2d4-372cdbededa9", "address": "fa:16:3e:6d:e6:cd", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0496010d-b0", "ovs_interfaceid": "0496010d-b03c-4002-b2d4-372cdbededa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.657135] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:e6:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0496010d-b03c-4002-b2d4-372cdbededa9', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.667061] env[61995]: DEBUG oslo.service.loopingcall [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.667513] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 917.667738] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-704c6209-92bf-4c9a-84b0-6ca0dc363279 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.692302] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.692302] env[61995]: value = "task-794840" [ 917.692302] env[61995]: _type = "Task" [ 917.692302] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.701545] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794840, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.738944] env[61995]: DEBUG nova.scheduler.client.report [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.779542] env[61995]: DEBUG nova.compute.manager [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Received event network-changed-0496010d-b03c-4002-b2d4-372cdbededa9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.779867] env[61995]: DEBUG nova.compute.manager [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Refreshing instance network info cache due to event network-changed-0496010d-b03c-4002-b2d4-372cdbededa9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 917.779969] env[61995]: DEBUG oslo_concurrency.lockutils [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] Acquiring lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.780133] env[61995]: DEBUG oslo_concurrency.lockutils [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] Acquired lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.780327] env[61995]: DEBUG nova.network.neutron [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Refreshing network info cache for port 0496010d-b03c-4002-b2d4-372cdbededa9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 918.007853] env[61995]: DEBUG oslo_concurrency.lockutils [req-ed93c226-0548-4cf5-bcee-90d3b38cc75f req-5629864a-cf12-4c44-a1a4-38f30ccc5e57 service nova] Releasing lock "refresh_cache-7d987006-afac-4973-ae69-f15b4752469b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.204115] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794840, 'name': CreateVM_Task, 'duration_secs': 0.415699} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.204115] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 918.204679] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.204851] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.205279] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.205901] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43adc297-3a6a-4c86-ad7e-61c8e7d0e5ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.211154] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 918.211154] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52884261-51d1-7eab-f725-c7ed0b1c0ea0" [ 918.211154] env[61995]: _type = "Task" [ 918.211154] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.219293] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52884261-51d1-7eab-f725-c7ed0b1c0ea0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.244591] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.245333] env[61995]: DEBUG nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.249320] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 4.587s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.512939] env[61995]: DEBUG nova.network.neutron [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updated VIF entry in instance network info cache for port 0496010d-b03c-4002-b2d4-372cdbededa9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.513482] env[61995]: DEBUG nova.network.neutron [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance_info_cache with network_info: [{"id": "0496010d-b03c-4002-b2d4-372cdbededa9", "address": "fa:16:3e:6d:e6:cd", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0496010d-b0", "ovs_interfaceid": "0496010d-b03c-4002-b2d4-372cdbededa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.722562] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52884261-51d1-7eab-f725-c7ed0b1c0ea0, 'name': SearchDatastore_Task, 'duration_secs': 0.021701} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.722833] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.723090] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.723343] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.723495] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.723680] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.723953] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fb2baa3-9b95-46bd-be4e-72a4aa87e7eb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.738718] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.738905] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 918.739720] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84ebdf33-4dc9-4795-8c6e-a42a9b36c78c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.745913] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 918.745913] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5269f309-1e90-7d97-4787-403b68800e78" [ 918.745913] env[61995]: _type = "Task" [ 918.745913] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.755647] env[61995]: DEBUG nova.compute.utils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.757159] env[61995]: DEBUG nova.objects.instance [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lazy-loading 'migration_context' on Instance uuid 4b49e7b1-7fd0-4e59-af79-5c0898967c35 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.758143] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5269f309-1e90-7d97-4787-403b68800e78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.758745] env[61995]: DEBUG nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.758907] env[61995]: DEBUG nova.network.neutron [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.798481] env[61995]: DEBUG nova.policy [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3efdbd0f8e0340dc8e4bfbf672d62f0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83064cbae17c429d8d084837635486da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 919.016797] env[61995]: DEBUG oslo_concurrency.lockutils [req-7664eb7b-2d5a-4602-81a6-55cef0047088 req-373ca5eb-3b13-48f0-83c9-4c100300b743 service nova] Releasing lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.142259] env[61995]: DEBUG nova.network.neutron [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Successfully created port: a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.257475] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5269f309-1e90-7d97-4787-403b68800e78, 'name': SearchDatastore_Task, 'duration_secs': 0.015156} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.258496] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ec49b97-647e-4e7a-b735-52acccda950d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.261583] env[61995]: DEBUG nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.271117] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 919.271117] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523c6f86-7642-5bab-03c9-b0adfca226e1" [ 919.271117] env[61995]: _type = "Task" [ 919.271117] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.279751] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523c6f86-7642-5bab-03c9-b0adfca226e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.555920] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172eef8a-b6f5-4387-8ee0-b4210deffef9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.564627] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f14a08b-d03e-4dc7-9034-828847166122 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.596110] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9417bee6-1ef9-44a5-8f41-fffc72b214c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.604661] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081e487a-a0ea-4623-9616-dad9bad3eefc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.620360] env[61995]: DEBUG nova.compute.provider_tree [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.781681] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523c6f86-7642-5bab-03c9-b0adfca226e1, 'name': SearchDatastore_Task, 'duration_secs': 0.016277} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.781911] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.782194] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 057ba9d6-4ff2-4da1-be7d-ecfadf738d74/057ba9d6-4ff2-4da1-be7d-ecfadf738d74.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 919.782462] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34737143-9fb4-4baf-ab43-d6883e727a9c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.790776] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 919.790776] env[61995]: value = "task-794841" [ 919.790776] env[61995]: _type = "Task" [ 919.790776] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.799344] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794841, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.123792] env[61995]: DEBUG nova.scheduler.client.report [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.273291] env[61995]: DEBUG nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.300418] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.300743] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.300913] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.301127] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.301298] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.301468] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.301692] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.301862] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.302121] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.302319] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.302502] env[61995]: DEBUG nova.virt.hardware [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.303558] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa894f22-29e0-4669-b1f6-784c33655216 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.310453] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794841, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.317872] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0a4563-e8c1-472a-a865-3f8f6f558758 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.718122] env[61995]: DEBUG nova.compute.manager [req-52487cc0-6772-485a-aa33-2795cb811ce0 req-24d9df9f-3eb2-4ae2-bbf6-38540934bf8b service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Received event network-vif-plugged-a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.718435] env[61995]: DEBUG oslo_concurrency.lockutils [req-52487cc0-6772-485a-aa33-2795cb811ce0 req-24d9df9f-3eb2-4ae2-bbf6-38540934bf8b service nova] Acquiring lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.718586] env[61995]: DEBUG oslo_concurrency.lockutils [req-52487cc0-6772-485a-aa33-2795cb811ce0 req-24d9df9f-3eb2-4ae2-bbf6-38540934bf8b service nova] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.718811] env[61995]: DEBUG oslo_concurrency.lockutils [req-52487cc0-6772-485a-aa33-2795cb811ce0 req-24d9df9f-3eb2-4ae2-bbf6-38540934bf8b service nova] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.719041] env[61995]: DEBUG nova.compute.manager [req-52487cc0-6772-485a-aa33-2795cb811ce0 req-24d9df9f-3eb2-4ae2-bbf6-38540934bf8b service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] No waiting events found dispatching network-vif-plugged-a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.719300] env[61995]: WARNING nova.compute.manager [req-52487cc0-6772-485a-aa33-2795cb811ce0 req-24d9df9f-3eb2-4ae2-bbf6-38540934bf8b service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Received unexpected event network-vif-plugged-a0291e27-4d18-41af-ba79-3d058998150e for instance with vm_state building and task_state spawning. [ 920.804139] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794841, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.792378} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.804442] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 057ba9d6-4ff2-4da1-be7d-ecfadf738d74/057ba9d6-4ff2-4da1-be7d-ecfadf738d74.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 920.804660] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.804930] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6045725-47e7-4a83-b6d5-b107a388461b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.813162] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 920.813162] env[61995]: value = "task-794842" [ 920.813162] env[61995]: _type = "Task" [ 920.813162] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.823543] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.839572] env[61995]: DEBUG nova.network.neutron [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Successfully updated port: a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.137277] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.888s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.143386] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.006s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.143640] env[61995]: DEBUG nova.objects.instance [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'resources' on Instance uuid a8b7bcd6-4753-4e1f-8b0b-777c882b99f5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.325575] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.244251} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.325850] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.326690] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fb71b9-8aa9-4489-88a8-b68dd40decf4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.350381] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 057ba9d6-4ff2-4da1-be7d-ecfadf738d74/057ba9d6-4ff2-4da1-be7d-ecfadf738d74.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.351030] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.351173] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.351322] env[61995]: DEBUG nova.network.neutron [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 921.353043] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-121ba3c4-4f3c-4bbb-add2-e74b1de32a9a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.376135] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 921.376135] env[61995]: value = "task-794843" [ 921.376135] env[61995]: _type = "Task" [ 921.376135] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.385448] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794843, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.888047] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794843, 'name': ReconfigVM_Task, 'duration_secs': 0.333636} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.888837] env[61995]: DEBUG nova.network.neutron [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.890715] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 057ba9d6-4ff2-4da1-be7d-ecfadf738d74/057ba9d6-4ff2-4da1-be7d-ecfadf738d74.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.891484] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-404e36b9-c7dc-493d-9346-fad159f65490 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.900222] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 921.900222] env[61995]: value = "task-794844" [ 921.900222] env[61995]: _type = "Task" [ 921.900222] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.909416] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794844, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.946772] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b2daf9-734d-4802-a2c1-965cb6a442c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.955174] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87f4987-be42-4490-be7f-26eb0f2cbc5b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.990998] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffcf470-9b95-4498-8c37-a011a668fc90 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.999784] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6546ea1-461e-44f4-870b-dee062aca4a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.016096] env[61995]: DEBUG nova.compute.provider_tree [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.067924] env[61995]: DEBUG nova.network.neutron [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updating instance_info_cache with network_info: [{"id": "a0291e27-4d18-41af-ba79-3d058998150e", "address": "fa:16:3e:2b:4f:84", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0291e27-4d", "ovs_interfaceid": "a0291e27-4d18-41af-ba79-3d058998150e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.410678] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794844, 'name': Rename_Task, 'duration_secs': 0.205064} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.412212] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 922.412212] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86f76766-0b65-4714-a8c1-72f1b6973031 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.419662] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 922.419662] env[61995]: value = "task-794845" [ 922.419662] env[61995]: _type = "Task" [ 922.419662] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.430095] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794845, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.519183] env[61995]: DEBUG nova.scheduler.client.report [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.570653] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.570980] env[61995]: DEBUG nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Instance network_info: |[{"id": "a0291e27-4d18-41af-ba79-3d058998150e", "address": "fa:16:3e:2b:4f:84", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0291e27-4d", "ovs_interfaceid": "a0291e27-4d18-41af-ba79-3d058998150e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.571442] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:4f:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0291e27-4d18-41af-ba79-3d058998150e', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.578838] env[61995]: DEBUG oslo.service.loopingcall [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.579090] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.579368] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5160f9bf-639d-48cd-8811-cbffce7a1266 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.600932] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.600932] env[61995]: value = "task-794846" [ 922.600932] env[61995]: _type = "Task" [ 922.600932] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.609788] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794846, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.682816] env[61995]: INFO nova.compute.manager [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Swapping old allocation on dict_keys(['5c086f4d-bc91-4e49-9831-bed8df133c15']) held by migration 4d70102d-a1be-41ff-88ca-3e9bb403b81b for instance [ 922.709240] env[61995]: DEBUG nova.scheduler.client.report [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Overwriting current allocation {'allocations': {'5c086f4d-bc91-4e49-9831-bed8df133c15': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 95}}, 'project_id': '5a7fbb0e40964f349256c7529874ca8e', 'user_id': 'c135f484cfe0401f83d1fcf16debf8b5', 'consumer_generation': 1} on consumer 4b49e7b1-7fd0-4e59-af79-5c0898967c35 {{(pid=61995) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 922.764465] env[61995]: DEBUG nova.compute.manager [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Received event network-changed-a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.764659] env[61995]: DEBUG nova.compute.manager [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Refreshing instance network info cache due to event network-changed-a0291e27-4d18-41af-ba79-3d058998150e. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.764893] env[61995]: DEBUG oslo_concurrency.lockutils [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] Acquiring lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.765055] env[61995]: DEBUG oslo_concurrency.lockutils [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] Acquired lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.765348] env[61995]: DEBUG nova.network.neutron [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Refreshing network info cache for port a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.804889] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.805187] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquired lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.805418] env[61995]: DEBUG nova.network.neutron [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.910249] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525b8c81-49f3-b5e2-be11-e81c3bde3e58/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 922.910948] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c756e58-547f-4dbf-842b-dc381a060660 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.920828] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525b8c81-49f3-b5e2-be11-e81c3bde3e58/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 922.921117] env[61995]: ERROR oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525b8c81-49f3-b5e2-be11-e81c3bde3e58/disk-0.vmdk due to incomplete transfer. [ 922.925338] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9040ce01-89e1-4d15-a081-1b8c6a662dbd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.933754] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794845, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.934925] env[61995]: DEBUG oslo_vmware.rw_handles [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525b8c81-49f3-b5e2-be11-e81c3bde3e58/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 922.935155] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Uploaded image 4c8c4eb7-8495-4e66-b897-ea812737ebc8 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 922.938058] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 922.938058] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e0767f2e-aa87-464b-a0e7-e94eeec65d6f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.946276] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 922.946276] env[61995]: value = "task-794847" [ 922.946276] env[61995]: _type = "Task" [ 922.946276] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.955431] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794847, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.024646] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.881s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.027354] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.017s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.027642] env[61995]: DEBUG nova.objects.instance [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lazy-loading 'resources' on Instance uuid fcdc7c33-8c46-4f4d-bb39-e6f943256cb0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.047124] env[61995]: INFO nova.scheduler.client.report [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted allocations for instance a8b7bcd6-4753-4e1f-8b0b-777c882b99f5 [ 923.113563] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794846, 'name': CreateVM_Task, 'duration_secs': 0.485819} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.113745] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 923.114478] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.114659] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.115083] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.115406] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b725a3b-423e-4629-8ae1-697066a634cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.121403] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 923.121403] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52996ead-fbca-f6c9-a457-fa37dff4335c" [ 923.121403] env[61995]: _type = "Task" [ 923.121403] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.130194] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52996ead-fbca-f6c9-a457-fa37dff4335c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.429941] env[61995]: DEBUG oslo_vmware.api [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794845, 'name': PowerOnVM_Task, 'duration_secs': 0.583786} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.432246] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 923.432464] env[61995]: INFO nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Took 8.04 seconds to spawn the instance on the hypervisor. [ 923.432651] env[61995]: DEBUG nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.433424] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c99f2d-2efb-4733-ba04-4a509fbb8560 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.455381] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794847, 'name': Destroy_Task, 'duration_secs': 0.428742} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.457883] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Destroyed the VM [ 923.458207] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 923.458459] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-25aad8d0-bada-4ab4-8100-724736193f2e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.465869] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 923.465869] env[61995]: value = "task-794848" [ 923.465869] env[61995]: _type = "Task" [ 923.465869] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.475215] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794848, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.555582] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a84d6a9-c53f-4124-a0c3-fe1f0af84d01 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "a8b7bcd6-4753-4e1f-8b0b-777c882b99f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.667s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.574244] env[61995]: DEBUG nova.network.neutron [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updated VIF entry in instance network info cache for port a0291e27-4d18-41af-ba79-3d058998150e. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.574244] env[61995]: DEBUG nova.network.neutron [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updating instance_info_cache with network_info: [{"id": "a0291e27-4d18-41af-ba79-3d058998150e", "address": "fa:16:3e:2b:4f:84", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0291e27-4d", "ovs_interfaceid": "a0291e27-4d18-41af-ba79-3d058998150e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.611691] env[61995]: DEBUG nova.network.neutron [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance_info_cache with network_info: [{"id": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "address": "fa:16:3e:45:26:f9", "network": {"id": "a8cf9779-532e-43e0-ab5f-8816c790232a", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ae745f566b90403c8e615c7069e2827b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb488f7-74", "ovs_interfaceid": "9eb488f7-7474-4b91-9b01-2f4378f5daca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.631740] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52996ead-fbca-f6c9-a457-fa37dff4335c, 'name': SearchDatastore_Task, 'duration_secs': 0.020163} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.634159] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.634406] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.634645] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.634802] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.634979] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.635620] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64c7f8f5-119b-4568-8cd6-87a569013699 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.646957] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.647160] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.647858] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4052b9e3-73cd-41a3-b34f-666d4ffd248c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.653974] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 923.653974] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525b2f7f-b6d8-3155-222c-5722a220e703" [ 923.653974] env[61995]: _type = "Task" [ 923.653974] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.665446] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525b2f7f-b6d8-3155-222c-5722a220e703, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.772817] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8c685c-9916-4d7e-b808-a156722a63e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.780606] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ea8340-f828-4fc6-805d-e3206aad87d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.813332] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614b2a12-7f36-4fd5-b82f-13f8446e490b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.822241] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4438f7a-8000-4723-9287-46ca9ef14258 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.836780] env[61995]: DEBUG nova.compute.provider_tree [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.953904] env[61995]: INFO nova.compute.manager [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Took 13.51 seconds to build instance. [ 923.978246] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794848, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.077518] env[61995]: DEBUG oslo_concurrency.lockutils [req-94a8ffe1-ec4b-4fea-99f2-b464811cd11c req-b21a9503-a325-410c-af69-40d5bcfaed8c service nova] Releasing lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.114923] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Releasing lock "refresh_cache-4b49e7b1-7fd0-4e59-af79-5c0898967c35" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.115440] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 924.115725] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49bb2768-2191-4116-aa1b-01221e1a8221 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.123825] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 924.123825] env[61995]: value = "task-794849" [ 924.123825] env[61995]: _type = "Task" [ 924.123825] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.133272] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.164990] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525b2f7f-b6d8-3155-222c-5722a220e703, 'name': SearchDatastore_Task, 'duration_secs': 0.019266} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.165875] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01794ae6-d544-407d-83c2-2752ab07fb10 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.172302] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 924.172302] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52de1c37-5a93-eb85-d8a8-c7a8a3380e82" [ 924.172302] env[61995]: _type = "Task" [ 924.172302] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.182734] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52de1c37-5a93-eb85-d8a8-c7a8a3380e82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.340072] env[61995]: DEBUG nova.scheduler.client.report [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.455651] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2bf23299-d18f-42a6-82be-78d778293afd tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.017s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.479297] env[61995]: DEBUG oslo_vmware.api [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794848, 'name': RemoveSnapshot_Task, 'duration_secs': 0.560741} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.479566] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 924.479802] env[61995]: INFO nova.compute.manager [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Took 15.29 seconds to snapshot the instance on the hypervisor. [ 924.634760] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794849, 'name': PowerOffVM_Task, 'duration_secs': 0.274157} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.635049] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 924.635715] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:58:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2ec0e7a2-ad31-4b6b-b0cd-3d774a987d95',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-892104472',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.635944] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.636387] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.636627] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.636796] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.636951] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.637181] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.637353] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.637527] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.637693] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.637869] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.643123] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd003b1b-746c-44fb-8e12-9191a4502265 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.661804] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 924.661804] env[61995]: value = "task-794850" [ 924.661804] env[61995]: _type = "Task" [ 924.661804] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.670838] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794850, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.684602] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52de1c37-5a93-eb85-d8a8-c7a8a3380e82, 'name': SearchDatastore_Task, 'duration_secs': 0.020132} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.684602] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.684826] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.685101] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39a8ad2a-3a3a-4d81-87d0-977a6eefcc0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.693514] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 924.693514] env[61995]: value = "task-794851" [ 924.693514] env[61995]: _type = "Task" [ 924.693514] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.706595] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.793121] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "88638f35-950d-42b4-92b2-ed0dced9148c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.793121] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "88638f35-950d-42b4-92b2-ed0dced9148c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.825561] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 924.825753] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 924.845345] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.818s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.847828] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.492s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.848098] env[61995]: DEBUG nova.objects.instance [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lazy-loading 'pci_requests' on Instance uuid 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.864062] env[61995]: INFO nova.scheduler.client.report [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted allocations for instance fcdc7c33-8c46-4f4d-bb39-e6f943256cb0 [ 925.043606] env[61995]: DEBUG nova.compute.manager [None req-59fa2a5c-eecc-4376-9236-8e6c2e58637c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Found 1 images (rotation: 2) {{(pid=61995) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 925.173988] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794850, 'name': ReconfigVM_Task, 'duration_secs': 0.210803} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.175396] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecee6ecb-3ac4-49c4-9ce5-2cd6e270d2b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.209445] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:58:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2ec0e7a2-ad31-4b6b-b0cd-3d774a987d95',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-892104472',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.210068] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.210212] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.210419] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.210571] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.210727] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.210948] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.211140] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.211324] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.211497] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.211681] env[61995]: DEBUG nova.virt.hardware [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.215581] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03d28de9-838a-4996-ac50-8bcb3b4815bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.224673] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 925.224673] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]524acb2e-e5b9-3689-c111-a109b3498599" [ 925.224673] env[61995]: _type = "Task" [ 925.224673] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.228195] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794851, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.240228] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524acb2e-e5b9-3689-c111-a109b3498599, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.295149] env[61995]: DEBUG nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 925.337028] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.337028] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 925.352938] env[61995]: DEBUG nova.objects.instance [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lazy-loading 'numa_topology' on Instance uuid 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.374835] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0e1bbfbb-c18e-4011-bffe-254d6e8e1a30 tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "fcdc7c33-8c46-4f4d-bb39-e6f943256cb0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.423s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.718741] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.812827} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.719640] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.720113] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.720449] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ab1120d-def5-416d-988f-dc6aa1804cf2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.728505] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 925.728505] env[61995]: value = "task-794852" [ 925.728505] env[61995]: _type = "Task" [ 925.728505] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.740846] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794852, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.744464] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524acb2e-e5b9-3689-c111-a109b3498599, 'name': SearchDatastore_Task, 'duration_secs': 0.028799} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.750472] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfiguring VM instance instance-00000043 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 925.750798] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb5f7fac-bc11-4dd4-b40e-177e56688397 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.770598] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 925.770598] env[61995]: value = "task-794853" [ 925.770598] env[61995]: _type = "Task" [ 925.770598] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.780413] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794853, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.821814] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.855961] env[61995]: INFO nova.compute.claims [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.871308] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.871507] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.871721] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 926.166442] env[61995]: DEBUG nova.compute.manager [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 926.241490] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794852, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066057} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.241768] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.242584] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a46868-b465-429f-8288-677d02696938 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.265015] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.265285] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08866a2e-6d6b-4dcc-bb56-06b27dfbbf14 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.288204] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794853, 'name': ReconfigVM_Task, 'duration_secs': 0.206111} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.289358] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfigured VM instance instance-00000043 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 926.289667] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 926.289667] env[61995]: value = "task-794854" [ 926.289667] env[61995]: _type = "Task" [ 926.289667] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.290348] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dec8641-4d9d-4493-989b-801977527479 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.300310] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794854, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.310614] env[61995]: DEBUG nova.compute.manager [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.318652] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.319599] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f765a8fb-01fe-4c4d-9b89-e959994b68c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.322226] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f409b114-16d2-49cb-a841-3b16d36b7fa7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.343572] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 926.343572] env[61995]: value = "task-794855" [ 926.343572] env[61995]: _type = "Task" [ 926.343572] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.354152] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.386062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.386062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.386062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.386062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.386062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.388632] env[61995]: INFO nova.compute.manager [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Terminating instance [ 926.391727] env[61995]: DEBUG nova.compute.manager [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.392249] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 926.392843] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cacebc8b-37ee-45ac-908c-f931b5a1449f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.402826] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.403159] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6f389c7-e7ba-4663-b382-e18fa63805cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.410506] env[61995]: DEBUG oslo_vmware.api [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 926.410506] env[61995]: value = "task-794856" [ 926.410506] env[61995]: _type = "Task" [ 926.410506] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.420373] env[61995]: DEBUG oslo_vmware.api [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.686149] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.803010] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794854, 'name': ReconfigVM_Task, 'duration_secs': 0.339721} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.803335] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfigured VM instance instance-0000004c to attach disk [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.803969] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd4bb828-7d57-4685-8fb9-9c0e90da0544 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.811430] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 926.811430] env[61995]: value = "task-794857" [ 926.811430] env[61995]: _type = "Task" [ 926.811430] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.821518] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794857, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.846106] env[61995]: INFO nova.compute.manager [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] instance snapshotting [ 926.846570] env[61995]: DEBUG nova.objects.instance [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'flavor' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.857528] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794855, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.925474] env[61995]: DEBUG oslo_vmware.api [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794856, 'name': PowerOffVM_Task, 'duration_secs': 0.322813} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.936963] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 926.936963] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 926.936963] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-555946c7-b39a-4733-84d0-de5839c6b62b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.028928] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.029202] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.029389] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleting the datastore file [datastore2] c4f744f9-36f3-4913-9ced-f0db93d3b4df {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.029718] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68323c79-197e-4960-99be-bf898ac58ed8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.040506] env[61995]: DEBUG oslo_vmware.api [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for the task: (returnval){ [ 927.040506] env[61995]: value = "task-794859" [ 927.040506] env[61995]: _type = "Task" [ 927.040506] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.050758] env[61995]: DEBUG oslo_vmware.api [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794859, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.210119] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933b41c4-98e7-40b9-9b23-a3212f3fe42f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.218657] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94eb3bb3-65da-42a9-8e71-2d5f234227ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.249960] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.251742] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1373e90-595c-486b-9d58-575a2796fa82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.260717] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4888c6b0-d1cb-4f65-be1e-560cb8200895 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.276427] env[61995]: DEBUG nova.compute.provider_tree [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.324170] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794857, 'name': Rename_Task, 'duration_secs': 0.177041} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.324473] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.324721] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b89c2290-15db-4054-af3b-9df7d21155b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.331659] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 927.331659] env[61995]: value = "task-794860" [ 927.331659] env[61995]: _type = "Task" [ 927.331659] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.339330] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794860, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.355037] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fef39e-deea-4090-b9c8-756878a8521a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.361323] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794855, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.377697] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf43758-6bd9-4e7e-98e0-b65b3510383c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.550680] env[61995]: DEBUG oslo_vmware.api [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Task: {'id': task-794859, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.380419} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.550906] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.551119] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 927.551307] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 927.551491] env[61995]: INFO nova.compute.manager [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Took 1.16 seconds to destroy the instance on the hypervisor. [ 927.551786] env[61995]: DEBUG oslo.service.loopingcall [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.552012] env[61995]: DEBUG nova.compute.manager [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.552135] env[61995]: DEBUG nova.network.neutron [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.755797] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.756221] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 927.756528] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.756752] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.756933] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.757137] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.757311] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.757494] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.757664] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 927.757841] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.780033] env[61995]: DEBUG nova.scheduler.client.report [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.842767] env[61995]: DEBUG oslo_vmware.api [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794860, 'name': PowerOnVM_Task, 'duration_secs': 0.485861} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.843703] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 927.844024] env[61995]: INFO nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Took 7.57 seconds to spawn the instance on the hypervisor. [ 927.844246] env[61995]: DEBUG nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.845071] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc938cc6-bb58-4468-a0a6-f6ad14d8c5c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.863698] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794855, 'name': ReconfigVM_Task, 'duration_secs': 1.305173} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.864014] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35/4b49e7b1-7fd0-4e59-af79-5c0898967c35.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.864879] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc05ae33-80ef-42cf-bbd1-5705af99270b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.886122] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6448fc02-2d91-4d88-8e4f-3df95843014b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.906885] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 927.908050] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ed9ddbf4-8c78-4124-a9bb-a52876b83601 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.910448] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2a6c02-311a-4ca8-b60d-8398ed9ba6bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.932079] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9d9d76-ad37-4f35-94ad-b5094b1b83e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.937264] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 927.937264] env[61995]: value = "task-794861" [ 927.937264] env[61995]: _type = "Task" [ 927.937264] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.948844] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.952139] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56f15e45-83a7-4569-88f9-bfe1e22f0200 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.953624] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794861, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.960366] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 927.960366] env[61995]: value = "task-794862" [ 927.960366] env[61995]: _type = "Task" [ 927.960366] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.969462] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794862, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.059848] env[61995]: DEBUG nova.compute.manager [req-709ea587-25c7-4505-a36a-4b609ab4589d req-73a2b372-c561-4d02-99de-81b9834fa94a service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Received event network-vif-deleted-22b10d97-e8c2-4f19-b6b6-d5a55fe432c8 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.059974] env[61995]: INFO nova.compute.manager [req-709ea587-25c7-4505-a36a-4b609ab4589d req-73a2b372-c561-4d02-99de-81b9834fa94a service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Neutron deleted interface 22b10d97-e8c2-4f19-b6b6-d5a55fe432c8; detaching it from the instance and deleting it from the info cache [ 928.060147] env[61995]: DEBUG nova.network.neutron [req-709ea587-25c7-4505-a36a-4b609ab4589d req-73a2b372-c561-4d02-99de-81b9834fa94a service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.261308] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.285517] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.438s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.288473] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.920s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.288780] env[61995]: DEBUG nova.objects.instance [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'resources' on Instance uuid b1bd98d9-bd0f-4abd-a188-e5267ada4852 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.372490] env[61995]: INFO nova.compute.manager [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Took 15.39 seconds to build instance. [ 928.409141] env[61995]: INFO nova.network.neutron [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating port 6ee71b78-02c5-4f63-a764-d5f1bb61107c with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 928.450047] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794861, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.475600] env[61995]: DEBUG oslo_vmware.api [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794862, 'name': PowerOnVM_Task, 'duration_secs': 0.399441} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.475982] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 928.537438] env[61995]: DEBUG nova.network.neutron [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.562290] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e84d7b0d-3f68-4eac-9bad-780847a99d66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.572904] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537cd90d-ed58-466a-82bd-02add66785cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.609145] env[61995]: DEBUG nova.compute.manager [req-709ea587-25c7-4505-a36a-4b609ab4589d req-73a2b372-c561-4d02-99de-81b9834fa94a service nova] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Detach interface failed, port_id=22b10d97-e8c2-4f19-b6b6-d5a55fe432c8, reason: Instance c4f744f9-36f3-4913-9ced-f0db93d3b4df could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 928.873606] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5facb2a-9df0-4545-8045-a1dba620f4d0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.907s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.899976] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.900469] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.900720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.900931] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.901145] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.904154] env[61995]: INFO nova.compute.manager [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Terminating instance [ 928.906569] env[61995]: DEBUG nova.compute.manager [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.906883] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 928.907808] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865ffbf0-9990-478f-a068-2884c23eeb8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.920459] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 928.920741] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c3ca3fc-e040-4fc0-a4ad-f45582f7ed0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.930085] env[61995]: DEBUG oslo_vmware.api [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 928.930085] env[61995]: value = "task-794863" [ 928.930085] env[61995]: _type = "Task" [ 928.930085] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.938938] env[61995]: DEBUG oslo_vmware.api [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794863, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.950879] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794861, 'name': CreateSnapshot_Task, 'duration_secs': 0.648709} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.951891] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 928.952933] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a05b03a-524f-4c2e-8c23-4504358d9ce2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.041749] env[61995]: INFO nova.compute.manager [-] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Took 1.49 seconds to deallocate network for instance. [ 929.116189] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07db58a-5841-4aba-b751-37d704efbefd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.124617] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f301421d-bc2a-439a-8492-781bad92bcf8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.159814] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f06bcb-d3d9-4a11-9e8c-49a75eb652f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.168794] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3f6d11-d505-42af-a2aa-7dad56da94cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.184468] env[61995]: DEBUG nova.compute.provider_tree [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.442261] env[61995]: DEBUG oslo_vmware.api [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794863, 'name': PowerOffVM_Task, 'duration_secs': 0.315121} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.442261] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 929.442261] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 929.442261] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eafd409a-59fc-4f68-9c6f-f8a3fbd5df8b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.480461] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 929.480810] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fe4ed748-fecd-4e64-af2f-8be50d5d45be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.487054] env[61995]: INFO nova.compute.manager [None req-1f57ecfc-9eca-47c1-9eca-2efbdb661f13 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance to original state: 'active' [ 929.491256] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 929.491256] env[61995]: value = "task-794865" [ 929.491256] env[61995]: _type = "Task" [ 929.491256] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.501032] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794865, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.533334] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 929.533334] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 929.533334] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleting the datastore file [datastore2] a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.533334] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e1a357c-e40f-4a6c-a7dd-acd0ad87f745 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.542969] env[61995]: DEBUG oslo_vmware.api [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 929.542969] env[61995]: value = "task-794866" [ 929.542969] env[61995]: _type = "Task" [ 929.542969] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.548033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.553324] env[61995]: DEBUG oslo_vmware.api [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.687843] env[61995]: DEBUG nova.scheduler.client.report [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.008718] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794865, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.055999] env[61995]: DEBUG oslo_vmware.api [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.087228] env[61995]: DEBUG nova.compute.manager [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Received event network-changed-a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 930.087568] env[61995]: DEBUG nova.compute.manager [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Refreshing instance network info cache due to event network-changed-a0291e27-4d18-41af-ba79-3d058998150e. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 930.087801] env[61995]: DEBUG oslo_concurrency.lockutils [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] Acquiring lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.087875] env[61995]: DEBUG oslo_concurrency.lockutils [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] Acquired lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.088130] env[61995]: DEBUG nova.network.neutron [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Refreshing network info cache for port a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.150701] env[61995]: DEBUG oslo_concurrency.lockutils [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.150901] env[61995]: DEBUG oslo_concurrency.lockutils [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.151493] env[61995]: INFO nova.compute.manager [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Rebooting instance [ 930.192866] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.195256] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.374s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.196895] env[61995]: INFO nova.compute.claims [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.221878] env[61995]: INFO nova.scheduler.client.report [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted allocations for instance b1bd98d9-bd0f-4abd-a188-e5267ada4852 [ 930.508717] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794865, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.558166] env[61995]: DEBUG oslo_vmware.api [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.532817} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.558166] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.558166] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 930.558166] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 930.558166] env[61995]: INFO nova.compute.manager [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Took 1.65 seconds to destroy the instance on the hypervisor. [ 930.558166] env[61995]: DEBUG oslo.service.loopingcall [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.558166] env[61995]: DEBUG nova.compute.manager [-] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 930.558669] env[61995]: DEBUG nova.network.neutron [-] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 930.675672] env[61995]: DEBUG oslo_concurrency.lockutils [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.675894] env[61995]: DEBUG oslo_concurrency.lockutils [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.676091] env[61995]: DEBUG nova.network.neutron [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 930.729527] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9096f19-1de1-4691-9428-4c23d70ab0aa tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "b1bd98d9-bd0f-4abd-a188-e5267ada4852" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.725s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.850897] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.851122] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.851316] env[61995]: DEBUG nova.network.neutron [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 930.888587] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.888978] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.889309] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.891740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.891740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.892844] env[61995]: INFO nova.compute.manager [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Terminating instance [ 930.895488] env[61995]: DEBUG nova.compute.manager [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.896301] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 930.899421] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1827f7f0-cefd-4f9b-a855-cd71e7320014 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.906314] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 930.906661] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a09b176d-6087-4eaa-84c2-f9d005fdc5bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.913964] env[61995]: DEBUG oslo_vmware.api [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 930.913964] env[61995]: value = "task-794867" [ 930.913964] env[61995]: _type = "Task" [ 930.913964] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.925334] env[61995]: DEBUG oslo_vmware.api [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.007892] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794865, 'name': CloneVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.152261] env[61995]: DEBUG nova.network.neutron [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updated VIF entry in instance network info cache for port a0291e27-4d18-41af-ba79-3d058998150e. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.152694] env[61995]: DEBUG nova.network.neutron [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updating instance_info_cache with network_info: [{"id": "a0291e27-4d18-41af-ba79-3d058998150e", "address": "fa:16:3e:2b:4f:84", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.179", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0291e27-4d", "ovs_interfaceid": "a0291e27-4d18-41af-ba79-3d058998150e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.424872] env[61995]: DEBUG oslo_vmware.api [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794867, 'name': PowerOffVM_Task, 'duration_secs': 0.259835} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.427300] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 931.427482] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 931.427742] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e5574d7-c6ba-4292-bdb4-97ba9775e37a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.465819] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1ec355-a1a5-40f4-9397-faa5fc3de3ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.479887] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4610f40-8fe4-402f-af0b-5065bf4501bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.524379] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d96ecb9-2be8-4bef-bd16-6c34c682a508 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.528501] env[61995]: DEBUG nova.compute.manager [req-60533816-5d0c-45b5-b256-5bd22b332cb5 req-81b3f8d0-89a5-4e48-8351-f4ae088bcae1 service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Received event network-vif-deleted-af010e12-84c9-4a5d-b901-4f519c51ce79 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 931.528806] env[61995]: INFO nova.compute.manager [req-60533816-5d0c-45b5-b256-5bd22b332cb5 req-81b3f8d0-89a5-4e48-8351-f4ae088bcae1 service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Neutron deleted interface af010e12-84c9-4a5d-b901-4f519c51ce79; detaching it from the instance and deleting it from the info cache [ 931.529108] env[61995]: DEBUG nova.network.neutron [req-60533816-5d0c-45b5-b256-5bd22b332cb5 req-81b3f8d0-89a5-4e48-8351-f4ae088bcae1 service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.531638] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 931.531638] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 931.531638] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Deleting the datastore file [datastore1] 4b49e7b1-7fd0-4e59-af79-5c0898967c35 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.533043] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-313cae08-f5ec-4276-a2ac-5d99e20efb80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.540771] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794865, 'name': CloneVM_Task, 'duration_secs': 1.62006} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.543983] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Created linked-clone VM from snapshot [ 931.544342] env[61995]: DEBUG oslo_vmware.api [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 931.544342] env[61995]: value = "task-794869" [ 931.544342] env[61995]: _type = "Task" [ 931.544342] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.545413] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3c13fd-876e-4536-8983-e71e84b98fd0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.548979] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ca083a-935d-4fb1-97ba-baeac69216a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.564674] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Uploading image 91a96abd-c3b8-4135-bd90-9f82c94fdfcb {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 931.576926] env[61995]: DEBUG nova.compute.provider_tree [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.578369] env[61995]: DEBUG oslo_vmware.api [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794869, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.579460] env[61995]: DEBUG nova.network.neutron [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.607444] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 931.607444] env[61995]: value = "vm-185362" [ 931.607444] env[61995]: _type = "VirtualMachine" [ 931.607444] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 931.607985] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-978adf51-07ed-49b3-8f86-4cbd5fd90210 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.615817] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease: (returnval){ [ 931.615817] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52be7b4a-73fb-4656-a16f-d4509a6e9baf" [ 931.615817] env[61995]: _type = "HttpNfcLease" [ 931.615817] env[61995]: } obtained for exporting VM: (result){ [ 931.615817] env[61995]: value = "vm-185362" [ 931.615817] env[61995]: _type = "VirtualMachine" [ 931.615817] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 931.616256] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the lease: (returnval){ [ 931.616256] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52be7b4a-73fb-4656-a16f-d4509a6e9baf" [ 931.616256] env[61995]: _type = "HttpNfcLease" [ 931.616256] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 931.623767] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 931.623767] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52be7b4a-73fb-4656-a16f-d4509a6e9baf" [ 931.623767] env[61995]: _type = "HttpNfcLease" [ 931.623767] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 931.656217] env[61995]: DEBUG oslo_concurrency.lockutils [req-92a2417f-ec05-44a6-9e37-5db264661b84 req-089e57e1-8d6c-452f-8676-d9c45924fd04 service nova] Releasing lock "refresh_cache-bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.830230] env[61995]: DEBUG nova.network.neutron [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee71b78-02", "ovs_interfaceid": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.977681] env[61995]: DEBUG nova.network.neutron [-] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.032510] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdb6e482-8f4c-43f4-819d-6f6392c0fb20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.043658] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73782073-3d47-4d74-9ba2-528325198888 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.063444] env[61995]: DEBUG oslo_vmware.api [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794869, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155649} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.063715] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.063906] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 932.064109] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 932.064303] env[61995]: INFO nova.compute.manager [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Took 1.17 seconds to destroy the instance on the hypervisor. [ 932.064538] env[61995]: DEBUG oslo.service.loopingcall [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.064728] env[61995]: DEBUG nova.compute.manager [-] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 932.064823] env[61995]: DEBUG nova.network.neutron [-] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 932.080877] env[61995]: DEBUG nova.scheduler.client.report [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.084470] env[61995]: DEBUG nova.compute.manager [req-60533816-5d0c-45b5-b256-5bd22b332cb5 req-81b3f8d0-89a5-4e48-8351-f4ae088bcae1 service nova] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Detach interface failed, port_id=af010e12-84c9-4a5d-b901-4f519c51ce79, reason: Instance a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 932.085663] env[61995]: DEBUG oslo_concurrency.lockutils [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.086948] env[61995]: DEBUG nova.compute.manager [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.087807] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857f8d68-d570-4273-ae92-bf4588d6ea99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.127988] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 932.127988] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52be7b4a-73fb-4656-a16f-d4509a6e9baf" [ 932.127988] env[61995]: _type = "HttpNfcLease" [ 932.127988] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 932.127988] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 932.127988] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52be7b4a-73fb-4656-a16f-d4509a6e9baf" [ 932.127988] env[61995]: _type = "HttpNfcLease" [ 932.127988] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 932.127988] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97c8f6a-f2fb-4046-b924-5e600611c304 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.137468] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ff665d-f6a7-d542-4a5e-a8cbe9f8c9c4/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 932.137811] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ff665d-f6a7-d542-4a5e-a8cbe9f8c9c4/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 932.201528] env[61995]: DEBUG nova.compute.manager [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received event network-vif-plugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.203590] env[61995]: DEBUG oslo_concurrency.lockutils [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.203590] env[61995]: DEBUG oslo_concurrency.lockutils [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.203590] env[61995]: DEBUG oslo_concurrency.lockutils [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.203590] env[61995]: DEBUG nova.compute.manager [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] No waiting events found dispatching network-vif-plugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.203590] env[61995]: WARNING nova.compute.manager [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received unexpected event network-vif-plugged-6ee71b78-02c5-4f63-a764-d5f1bb61107c for instance with vm_state shelved_offloaded and task_state spawning. [ 932.203590] env[61995]: DEBUG nova.compute.manager [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received event network-changed-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.203590] env[61995]: DEBUG nova.compute.manager [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Refreshing instance network info cache due to event network-changed-6ee71b78-02c5-4f63-a764-d5f1bb61107c. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 932.203590] env[61995]: DEBUG oslo_concurrency.lockutils [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] Acquiring lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.303754] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1ac6e4ac-71e4-41df-8d52-25629d6fd99a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.337338] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.340030] env[61995]: DEBUG oslo_concurrency.lockutils [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] Acquired lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.340309] env[61995]: DEBUG nova.network.neutron [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Refreshing network info cache for port 6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.361346] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='cd16966bd8e7b2f6542822a32046e299',container_format='bare',created_at=2024-09-10T08:59:00Z,direct_url=,disk_format='vmdk',id=d9aa4058-3449-47b3-b10b-42d75d2fcc13,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1182386297-shelved',owner='3e0840f3a4d24ea78c35c4b273f1e413',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-09-10T08:59:17Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.361654] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.361826] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.362019] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.362177] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.362430] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.362542] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.362789] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.363086] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.363217] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.363423] env[61995]: DEBUG nova.virt.hardware [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.364275] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc02170b-a357-46d0-9186-7e3cad0bd550 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.373531] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98af3e6-5e52-4af6-a0b2-d58e7e6fb4fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.388300] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:b9:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff1f3320-df8e-49df-a412-9797a23bd173', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ee71b78-02c5-4f63-a764-d5f1bb61107c', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.397909] env[61995]: DEBUG oslo.service.loopingcall [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.397909] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 932.397909] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d2a175c-b6ff-4274-8da1-bc04e4c7bd72 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.418202] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.418202] env[61995]: value = "task-794871" [ 932.418202] env[61995]: _type = "Task" [ 932.418202] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.427462] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794871, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.481174] env[61995]: INFO nova.compute.manager [-] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Took 1.92 seconds to deallocate network for instance. [ 932.588399] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.589186] env[61995]: DEBUG nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.596183] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.907s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.920054] env[61995]: DEBUG nova.network.neutron [-] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.937843] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794871, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.989105] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.098921] env[61995]: DEBUG nova.compute.utils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.103739] env[61995]: INFO nova.compute.claims [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.107998] env[61995]: DEBUG nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.111350] env[61995]: DEBUG nova.network.neutron [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 933.116634] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656efa16-baab-4178-acfe-c27e4a712b82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.130108] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Doing hard reboot of VM {{(pid=61995) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 933.134042] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-44eb7ca9-ef05-4d9d-a6bf-68cdf2461977 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.140515] env[61995]: DEBUG oslo_vmware.api [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 933.140515] env[61995]: value = "task-794872" [ 933.140515] env[61995]: _type = "Task" [ 933.140515] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.152425] env[61995]: DEBUG oslo_vmware.api [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794872, 'name': ResetVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.190628] env[61995]: DEBUG nova.policy [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caf9bbbd848d4d7ba53ba3bd553943e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2018aae65cf4bf98b5ac82519201cf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 933.197146] env[61995]: DEBUG nova.network.neutron [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updated VIF entry in instance network info cache for port 6ee71b78-02c5-4f63-a764-d5f1bb61107c. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 933.197146] env[61995]: DEBUG nova.network.neutron [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee71b78-02", "ovs_interfaceid": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.430158] env[61995]: INFO nova.compute.manager [-] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Took 1.37 seconds to deallocate network for instance. [ 933.430483] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794871, 'name': CreateVM_Task, 'duration_secs': 0.560657} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.434960] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 933.439873] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.440193] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.440683] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.441398] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73d510a3-0052-4b58-ad2e-f2140b1867a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.448213] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 933.448213] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522ec49b-4376-c19c-8bca-9b97d454aba2" [ 933.448213] env[61995]: _type = "Task" [ 933.448213] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.457644] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522ec49b-4376-c19c-8bca-9b97d454aba2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.608355] env[61995]: DEBUG nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.614401] env[61995]: DEBUG nova.network.neutron [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Successfully created port: 272691df-c5ce-4116-a3e1-5e15186723b5 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.618481] env[61995]: INFO nova.compute.resource_tracker [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating resource usage from migration f6f8d556-5345-4a35-8742-c2e7db554c68 [ 933.654402] env[61995]: DEBUG oslo_vmware.api [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-794872, 'name': ResetVM_Task, 'duration_secs': 0.105542} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.654800] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Did hard reboot of VM {{(pid=61995) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 933.655067] env[61995]: DEBUG nova.compute.manager [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.655974] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799c4b78-f6b6-4f69-9a93-fbfd218cda44 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.703313] env[61995]: DEBUG oslo_concurrency.lockutils [req-d00cf55e-b1d4-4b2f-b54f-0684832fddad req-e9e14b51-76a2-45f8-81e0-b8c36c7b0ab3 service nova] Releasing lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.902432] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d479b519-7f00-4a48-8912-66880fb2149c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.911324] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3fc31b-59c7-4d7a-b0e2-8b3700fed47e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.945092] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.946148] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d639ff90-9495-47bf-a7d0-48b673c15ee7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.961300] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebea29d-220f-4390-b81a-e7484063ad02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.967132] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.967400] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Processing image d9aa4058-3449-47b3-b10b-42d75d2fcc13 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.967663] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13/d9aa4058-3449-47b3-b10b-42d75d2fcc13.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.967868] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13/d9aa4058-3449-47b3-b10b-42d75d2fcc13.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.968411] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.971782] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5712e8f3-6d89-4909-9f2a-d6fef868edb8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.973702] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.974036] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.983584] env[61995]: DEBUG nova.compute.provider_tree [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.992833] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.993065] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 933.994026] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27983cd3-0e6e-49f9-9b0c-53693cfa4cd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.001223] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 934.001223] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f4a14f-38a8-6b61-fd61-e17167f99a7d" [ 934.001223] env[61995]: _type = "Task" [ 934.001223] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.009621] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f4a14f-38a8-6b61-fd61-e17167f99a7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.176332] env[61995]: DEBUG oslo_concurrency.lockutils [None req-133aa95f-c089-4dd2-8245-1449aeb40584 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.025s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.270379] env[61995]: DEBUG nova.compute.manager [req-8c8a35db-245a-4a5c-bdf9-aa4eab9a0377 req-cd5fd18d-bbd8-489a-8018-769138db05c8 service nova] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Received event network-vif-deleted-9eb488f7-7474-4b91-9b01-2f4378f5daca {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.486435] env[61995]: DEBUG nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.491759] env[61995]: DEBUG nova.scheduler.client.report [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.513605] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Preparing fetch location {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 934.513892] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Fetch image to [datastore2] OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60/OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60.vmdk {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 934.514095] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Downloading stream optimized image d9aa4058-3449-47b3-b10b-42d75d2fcc13 to [datastore2] OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60/OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60.vmdk on the data store datastore2 as vApp {{(pid=61995) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 934.514274] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Downloading image file data d9aa4058-3449-47b3-b10b-42d75d2fcc13 to the ESX as VM named 'OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60' {{(pid=61995) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 934.598287] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 934.598287] env[61995]: value = "resgroup-9" [ 934.598287] env[61995]: _type = "ResourcePool" [ 934.598287] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 934.598587] env[61995]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-43cf578d-eba1-4b7e-a3d5-e638c9aa84b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.622875] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lease: (returnval){ [ 934.622875] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5268d599-4b99-ef35-2df0-23ef33e21cc2" [ 934.622875] env[61995]: _type = "HttpNfcLease" [ 934.622875] env[61995]: } obtained for vApp import into resource pool (val){ [ 934.622875] env[61995]: value = "resgroup-9" [ 934.622875] env[61995]: _type = "ResourcePool" [ 934.622875] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 934.623320] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the lease: (returnval){ [ 934.623320] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5268d599-4b99-ef35-2df0-23ef33e21cc2" [ 934.623320] env[61995]: _type = "HttpNfcLease" [ 934.623320] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 934.630607] env[61995]: DEBUG nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.636945] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 934.636945] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5268d599-4b99-ef35-2df0-23ef33e21cc2" [ 934.636945] env[61995]: _type = "HttpNfcLease" [ 934.636945] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 935.000378] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.407s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.000598] env[61995]: INFO nova.compute.manager [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Migrating [ 935.009656] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.749s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.009816] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.009974] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 935.010394] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.463s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.010653] env[61995]: DEBUG nova.objects.instance [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lazy-loading 'resources' on Instance uuid c4f744f9-36f3-4913-9ced-f0db93d3b4df {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.016409] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f8d768-29eb-40e7-b1fe-a252ee49bf38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.033928] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.041268] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ecdb48-29f3-437f-89ea-e17571384895 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.061874] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebca0b1d-8024-4eb1-ac86-d642179b57df {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.071902] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0ed2e8-6d7e-40a3-bf10-ab51760eb6df {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.109942] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179122MB free_disk=66GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 935.110239] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.114439] env[61995]: DEBUG nova.compute.manager [req-bd4f4bdb-9cc3-456f-b560-539e7fc1679b req-88d06f11-5015-4139-8933-baf1f980e358 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Received event network-vif-plugged-272691df-c5ce-4116-a3e1-5e15186723b5 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.114712] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd4f4bdb-9cc3-456f-b560-539e7fc1679b req-88d06f11-5015-4139-8933-baf1f980e358 service nova] Acquiring lock "88638f35-950d-42b4-92b2-ed0dced9148c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.114891] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd4f4bdb-9cc3-456f-b560-539e7fc1679b req-88d06f11-5015-4139-8933-baf1f980e358 service nova] Lock "88638f35-950d-42b4-92b2-ed0dced9148c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.115108] env[61995]: DEBUG oslo_concurrency.lockutils [req-bd4f4bdb-9cc3-456f-b560-539e7fc1679b req-88d06f11-5015-4139-8933-baf1f980e358 service nova] Lock "88638f35-950d-42b4-92b2-ed0dced9148c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.115319] env[61995]: DEBUG nova.compute.manager [req-bd4f4bdb-9cc3-456f-b560-539e7fc1679b req-88d06f11-5015-4139-8933-baf1f980e358 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] No waiting events found dispatching network-vif-plugged-272691df-c5ce-4116-a3e1-5e15186723b5 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.115533] env[61995]: WARNING nova.compute.manager [req-bd4f4bdb-9cc3-456f-b560-539e7fc1679b req-88d06f11-5015-4139-8933-baf1f980e358 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Received unexpected event network-vif-plugged-272691df-c5ce-4116-a3e1-5e15186723b5 for instance with vm_state building and task_state spawning. [ 935.138248] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 935.138248] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5268d599-4b99-ef35-2df0-23ef33e21cc2" [ 935.138248] env[61995]: _type = "HttpNfcLease" [ 935.138248] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 935.209522] env[61995]: DEBUG nova.network.neutron [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Successfully updated port: 272691df-c5ce-4116-a3e1-5e15186723b5 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.343252] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a5ddcc-5597-493b-923d-8f532f56fdda {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.351756] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c2ff5c-fc5f-46f6-8fff-70790d71b62c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.383523] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35345b12-d081-4d57-9501-052a56cb5493 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.392414] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b6225e-88e5-4338-b704-a313b93a3d34 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.406935] env[61995]: DEBUG nova.compute.provider_tree [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.529192] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.529405] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.529614] env[61995]: DEBUG nova.network.neutron [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.634046] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 935.634046] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5268d599-4b99-ef35-2df0-23ef33e21cc2" [ 935.634046] env[61995]: _type = "HttpNfcLease" [ 935.634046] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 935.634505] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 935.634505] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5268d599-4b99-ef35-2df0-23ef33e21cc2" [ 935.634505] env[61995]: _type = "HttpNfcLease" [ 935.634505] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 935.635141] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b83d43-22f9-480f-91ce-7b231fd0d534 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.645649] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b441-1f6f-5920-9a5a-25f42d035412/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 935.645971] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b441-1f6f-5920-9a5a-25f42d035412/disk-0.vmdk. {{(pid=61995) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 935.712047] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e34916ff-823a-4b83-9b14-6e722636924b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.715232] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "refresh_cache-88638f35-950d-42b4-92b2-ed0dced9148c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.715290] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "refresh_cache-88638f35-950d-42b4-92b2-ed0dced9148c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.715448] env[61995]: DEBUG nova.network.neutron [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.910336] env[61995]: DEBUG nova.scheduler.client.report [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.252034] env[61995]: DEBUG nova.network.neutron [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 936.260368] env[61995]: DEBUG nova.network.neutron [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance_info_cache with network_info: [{"id": "0496010d-b03c-4002-b2d4-372cdbededa9", "address": "fa:16:3e:6d:e6:cd", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0496010d-b0", "ovs_interfaceid": "0496010d-b03c-4002-b2d4-372cdbededa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.383191] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.383511] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.383679] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.383896] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.384027] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.384171] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.384429] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.384601] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.384803] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.384959] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.385145] env[61995]: DEBUG nova.virt.hardware [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.386082] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d26ce5-6685-4d04-9162-8fb932602f2a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.399195] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ac9055-84b3-4bcc-b0b2-e44481bc9d17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.417094] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.407s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.423765] env[61995]: DEBUG nova.network.neutron [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Updating instance_info_cache with network_info: [{"id": "272691df-c5ce-4116-a3e1-5e15186723b5", "address": "fa:16:3e:87:94:8e", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap272691df-c5", "ovs_interfaceid": "272691df-c5ce-4116-a3e1-5e15186723b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.424823] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.436s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.425099] env[61995]: DEBUG nova.objects.instance [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'resources' on Instance uuid a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.441675] env[61995]: INFO nova.scheduler.client.report [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Deleted allocations for instance c4f744f9-36f3-4913-9ced-f0db93d3b4df [ 936.764845] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.931661] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "refresh_cache-88638f35-950d-42b4-92b2-ed0dced9148c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.932163] env[61995]: DEBUG nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Instance network_info: |[{"id": "272691df-c5ce-4116-a3e1-5e15186723b5", "address": "fa:16:3e:87:94:8e", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap272691df-c5", "ovs_interfaceid": "272691df-c5ce-4116-a3e1-5e15186723b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.935330] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Completed reading data from the image iterator. {{(pid=61995) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 936.935679] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b441-1f6f-5920-9a5a-25f42d035412/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 936.936321] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:94:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '272691df-c5ce-4116-a3e1-5e15186723b5', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.945344] env[61995]: DEBUG oslo.service.loopingcall [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.946302] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1c06f7-2f80-4fa3-b474-12e2e3177426 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.949504] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 936.949788] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8aee3021-c88a-43ec-b5a3-5684729e46a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.971077] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1fde4da9-8dad-4ead-a9a6-65fa5c024f0d tempest-ServersAdminTestJSON-901921000 tempest-ServersAdminTestJSON-901921000-project-member] Lock "c4f744f9-36f3-4913-9ced-f0db93d3b4df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.585s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.974911] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b441-1f6f-5920-9a5a-25f42d035412/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 936.975204] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b441-1f6f-5920-9a5a-25f42d035412/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 936.975458] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-371606c9-7308-435f-ab89-4fef95f7f309 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.982041] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.982041] env[61995]: value = "task-794874" [ 936.982041] env[61995]: _type = "Task" [ 936.982041] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.995523] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794874, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.118216] env[61995]: INFO nova.compute.manager [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Rebuilding instance [ 937.142758] env[61995]: DEBUG nova.compute.manager [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Received event network-changed-272691df-c5ce-4116-a3e1-5e15186723b5 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 937.143053] env[61995]: DEBUG nova.compute.manager [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Refreshing instance network info cache due to event network-changed-272691df-c5ce-4116-a3e1-5e15186723b5. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 937.143303] env[61995]: DEBUG oslo_concurrency.lockutils [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] Acquiring lock "refresh_cache-88638f35-950d-42b4-92b2-ed0dced9148c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.143523] env[61995]: DEBUG oslo_concurrency.lockutils [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] Acquired lock "refresh_cache-88638f35-950d-42b4-92b2-ed0dced9148c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.143721] env[61995]: DEBUG nova.network.neutron [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Refreshing network info cache for port 272691df-c5ce-4116-a3e1-5e15186723b5 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.174616] env[61995]: DEBUG nova.compute.manager [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.176217] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac4809d-aa89-42ce-b25a-5b3b8dea691f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.234621] env[61995]: DEBUG oslo_vmware.rw_handles [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5235b441-1f6f-5920-9a5a-25f42d035412/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 937.234873] env[61995]: INFO nova.virt.vmwareapi.images [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Downloaded image file data d9aa4058-3449-47b3-b10b-42d75d2fcc13 [ 937.236986] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8da5c3-fae6-4f26-a21f-9a5dd00fcc1e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.242383] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec64a116-638b-4f41-a6cb-547d522174b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.262044] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ead471b-a2c7-4805-964e-10ba0b6b77db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.266491] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0eed7fa1-66ec-4494-b1e0-71301f82bbdf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.300171] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369f9b4b-3e97-43a6-9fc2-4cf340f1c881 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.310460] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762e33d5-c48d-4b55-8e8d-0586d1a0f5f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.327080] env[61995]: DEBUG nova.compute.provider_tree [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.329966] env[61995]: INFO nova.virt.vmwareapi.images [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] The imported VM was unregistered [ 937.331813] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Caching image {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 937.332041] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.332705] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97ca0c2f-b504-453a-b295-137a832768e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.348067] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Created directory with path [datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.348285] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60/OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60.vmdk to [datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13/d9aa4058-3449-47b3-b10b-42d75d2fcc13.vmdk. {{(pid=61995) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 937.348546] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ca818a2e-9285-4bf1-8f20-ca678dd99f48 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.357755] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 937.357755] env[61995]: value = "task-794876" [ 937.357755] env[61995]: _type = "Task" [ 937.357755] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.366211] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794876, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.495327] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794874, 'name': CreateVM_Task, 'duration_secs': 0.469247} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.495549] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 937.496378] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.496660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.496965] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.497311] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b262c584-df0b-406c-8ebb-7e8630551d42 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.503123] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 937.503123] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c9a265-9464-bad6-daff-d347828fe6bb" [ 937.503123] env[61995]: _type = "Task" [ 937.503123] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.512833] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c9a265-9464-bad6-daff-d347828fe6bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.693808] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 937.694212] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22e7644a-3087-46a2-ac28-238bb2d3f8b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.704391] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 937.704391] env[61995]: value = "task-794877" [ 937.704391] env[61995]: _type = "Task" [ 937.704391] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.721690] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.832492] env[61995]: DEBUG nova.scheduler.client.report [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.869945] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794876, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.987736] env[61995]: DEBUG nova.network.neutron [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Updated VIF entry in instance network info cache for port 272691df-c5ce-4116-a3e1-5e15186723b5. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 937.988172] env[61995]: DEBUG nova.network.neutron [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Updating instance_info_cache with network_info: [{"id": "272691df-c5ce-4116-a3e1-5e15186723b5", "address": "fa:16:3e:87:94:8e", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap272691df-c5", "ovs_interfaceid": "272691df-c5ce-4116-a3e1-5e15186723b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.016192] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c9a265-9464-bad6-daff-d347828fe6bb, 'name': SearchDatastore_Task, 'duration_secs': 0.01135} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.016722] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.017054] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.017321] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.017475] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.017663] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.018223] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ee4754b-8f44-4018-b662-1a63ac758342 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.037783] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.038147] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 938.040606] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e15f1194-6d4a-4125-84d8-4e7776ab9a7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.046511] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 938.046511] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52284c14-7973-d3a9-9c82-1e44eaa46c61" [ 938.046511] env[61995]: _type = "Task" [ 938.046511] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.056797] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52284c14-7973-d3a9-9c82-1e44eaa46c61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.219483] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794877, 'name': PowerOffVM_Task, 'duration_secs': 0.280118} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.220324] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.223046] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.223400] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76f463d5-5faa-42dc-9592-b350266cb602 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.236635] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 938.236635] env[61995]: value = "task-794878" [ 938.236635] env[61995]: _type = "Task" [ 938.236635] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.246711] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 938.246985] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 938.247256] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185324', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'name': 'volume-604ce120-96fa-4557-9558-055ca5966123', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ca0dcd7-a2b8-4de6-835b-07c36ab95a99', 'attached_at': '', 'detached_at': '', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'serial': '604ce120-96fa-4557-9558-055ca5966123'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 938.248214] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a71a13-3d5c-4965-adca-2cb2798ebb79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.272292] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70d36b2-eec9-4955-86e3-5072e62d14f3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.282686] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09939d22-eb83-4ea9-b897-cee75b039adf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.286047] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29fa3e6-78fe-48f6-9bd3-4d37cf2933a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.321779] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3de33c5-d7d2-40e3-9cc8-771aca706ffb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.324760] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance '057ba9d6-4ff2-4da1-be7d-ecfadf738d74' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 938.345948] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.921s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.347943] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] The volume has not been displaced from its original location: [datastore1] volume-604ce120-96fa-4557-9558-055ca5966123/volume-604ce120-96fa-4557-9558-055ca5966123.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 938.353543] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Reconfiguring VM instance instance-00000046 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 938.355096] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.409s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.355096] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.356967] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.323s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.358456] env[61995]: INFO nova.compute.claims [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.365136] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c2713fb-860c-402b-bc1e-9ca042a1ec65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.388892] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794876, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.390616] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 938.390616] env[61995]: value = "task-794879" [ 938.390616] env[61995]: _type = "Task" [ 938.390616] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.401710] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794879, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.403689] env[61995]: INFO nova.scheduler.client.report [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted allocations for instance a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5 [ 938.407091] env[61995]: INFO nova.scheduler.client.report [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Deleted allocations for instance 4b49e7b1-7fd0-4e59-af79-5c0898967c35 [ 938.498616] env[61995]: DEBUG oslo_concurrency.lockutils [req-a859a161-5185-444b-bf1e-c1eb20e1f981 req-c4aca99c-aabf-4f4d-b7e2-0f6f15251c30 service nova] Releasing lock "refresh_cache-88638f35-950d-42b4-92b2-ed0dced9148c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.562379] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52284c14-7973-d3a9-9c82-1e44eaa46c61, 'name': SearchDatastore_Task, 'duration_secs': 0.088857} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.562514] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bffba5dc-f687-4b46-aed7-38a1d6380de2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.572685] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 938.572685] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d25539-6d6f-03bc-2a90-d3c32292e57d" [ 938.572685] env[61995]: _type = "Task" [ 938.572685] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.585619] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d25539-6d6f-03bc-2a90-d3c32292e57d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.831401] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.832732] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2eb9e845-d160-472b-a33c-a1a2d552f11b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.848390] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 938.848390] env[61995]: value = "task-794880" [ 938.848390] env[61995]: _type = "Task" [ 938.848390] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.868064] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.893752] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794876, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.912894] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794879, 'name': ReconfigVM_Task, 'duration_secs': 0.264918} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.913596] env[61995]: DEBUG oslo_concurrency.lockutils [None req-22ee2337-8bb0-46d3-b016-97d1daac81c8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.013s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.915516] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Reconfigured VM instance instance-00000046 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 938.924956] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13268c14-9832-4a0e-8e72-2f879c0bdc5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.940217] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfc5e1cd-d468-427a-918c-923d51fb5287 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4b49e7b1-7fd0-4e59-af79-5c0898967c35" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.051s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.952528] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 938.952528] env[61995]: value = "task-794881" [ 938.952528] env[61995]: _type = "Task" [ 938.952528] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.965087] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.086883] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d25539-6d6f-03bc-2a90-d3c32292e57d, 'name': SearchDatastore_Task, 'duration_secs': 0.094739} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.088203] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.088203] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 88638f35-950d-42b4-92b2-ed0dced9148c/88638f35-950d-42b4-92b2-ed0dced9148c.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 939.088203] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65acbc24-5bca-4fbd-a3ce-0d371994cbe9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.100587] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 939.100587] env[61995]: value = "task-794882" [ 939.100587] env[61995]: _type = "Task" [ 939.100587] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.115338] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.363613] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794880, 'name': PowerOffVM_Task, 'duration_secs': 0.379779} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.363613] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.363613] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance '057ba9d6-4ff2-4da1-be7d-ecfadf738d74' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 939.391977] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794876, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.468434] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794881, 'name': ReconfigVM_Task, 'duration_secs': 0.271522} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.471977] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185324', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'name': 'volume-604ce120-96fa-4557-9558-055ca5966123', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ca0dcd7-a2b8-4de6-835b-07c36ab95a99', 'attached_at': '', 'detached_at': '', 'volume_id': '604ce120-96fa-4557-9558-055ca5966123', 'serial': '604ce120-96fa-4557-9558-055ca5966123'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 939.472647] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 939.474352] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6b7442-0e34-44fa-8e10-a9b9c96bb7d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.486264] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.488490] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cba5c330-3109-4be1-8a48-76b3a31f5d56 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.582557] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.582835] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.583040] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Deleting the datastore file [datastore1] 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.583344] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb4ba417-a92e-4162-bd4c-7678ed179b8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.593763] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for the task: (returnval){ [ 939.593763] env[61995]: value = "task-794884" [ 939.593763] env[61995]: _type = "Task" [ 939.593763] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.607923] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.617274] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.662844] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2047205-f5cb-4fbd-b502-933f02019eca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.674186] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42cec37-a032-42ab-91d0-9ce35838ae8e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.707430] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64851a1-5df7-4dcb-9dfa-4bf794c2cc3c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.716936] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af46222-e8a5-4a81-bc92-afde2934bb50 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.732998] env[61995]: DEBUG nova.compute.provider_tree [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.833544] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.833926] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.870478] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.870715] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.870884] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.871160] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.871358] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.871520] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.871737] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.871902] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.872438] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.872438] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.872610] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.879051] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3343ddb-144f-4617-831a-cf9ff661baf7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.903488] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794876, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.905559] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 939.905559] env[61995]: value = "task-794885" [ 939.905559] env[61995]: _type = "Task" [ 939.905559] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.916529] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794885, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.935065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.935375] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.107356] env[61995]: DEBUG oslo_vmware.api [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Task: {'id': task-794884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142375} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.111033] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.111335] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 940.111479] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 940.126760] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794882, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.195150] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 940.195150] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c2436d5-7426-4033-9fa5-a5db85d8c2f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.207499] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec077217-2df1-40ae-a181-4e11f11602a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.254986] env[61995]: DEBUG nova.scheduler.client.report [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.262597] env[61995]: ERROR nova.compute.manager [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Failed to detach volume 604ce120-96fa-4557-9558-055ca5966123 from /dev/sda: nova.exception.InstanceNotFound: Instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 could not be found. [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Traceback (most recent call last): [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self.driver.rebuild(**kwargs) [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/driver.py", line 392, in rebuild [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] raise NotImplementedError() [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] NotImplementedError [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] During handling of the above exception, another exception occurred: [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Traceback (most recent call last): [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self.driver.detach_volume(context, old_connection_info, [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] return self._volumeops.detach_volume(connection_info, instance) [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self._detach_volume_vmdk(connection_info, instance) [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] stable_ref.fetch_moref(session) [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] raise exception.InstanceNotFound(instance_id=self._uuid) [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] nova.exception.InstanceNotFound: Instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 could not be found. [ 940.262597] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.339950] env[61995]: DEBUG nova.compute.utils [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.403102] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794876, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.754432} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.404022] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60/OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60.vmdk to [datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13/d9aa4058-3449-47b3-b10b-42d75d2fcc13.vmdk. [ 940.404367] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Cleaning up location [datastore2] OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 940.404703] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_b72fb4b8-097c-46ef-9cda-454374cd4b60 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.405244] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f91b6e0-9ec5-4d8a-b0ef-4be291e5f9bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.420905] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794885, 'name': ReconfigVM_Task, 'duration_secs': 0.288057} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.423429] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance '057ba9d6-4ff2-4da1-be7d-ecfadf738d74' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 940.431261] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 940.431261] env[61995]: value = "task-794886" [ 940.431261] env[61995]: _type = "Task" [ 940.431261] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.439297] env[61995]: INFO nova.compute.manager [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Detaching volume 34345e8c-85a0-4c01-85e5-bc6ec70b8cc9 [ 940.450076] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794886, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.476964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.476964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.476964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.476964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.476964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.479228] env[61995]: INFO nova.compute.manager [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Terminating instance [ 940.483905] env[61995]: DEBUG nova.compute.manager [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 940.483905] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 940.483905] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c1307c-0932-43e5-b003-5fae066c5e7e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.491440] env[61995]: INFO nova.virt.block_device [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Attempting to driver detach volume 34345e8c-85a0-4c01-85e5-bc6ec70b8cc9 from mountpoint /dev/sdb [ 940.491880] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 940.492253] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185322', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'name': 'volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc7077ac-d3fd-4e84-867a-d86328b32f5d', 'attached_at': '', 'detached_at': '', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'serial': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 940.493064] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae41787-ef6b-4002-9cf6-8ebd7aff9513 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.500534] env[61995]: DEBUG nova.compute.utils [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Build of instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 aborted: Failed to rebuild volume backed instance. {{(pid=61995) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 940.501983] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.503182] env[61995]: ERROR nova.compute.manager [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 aborted: Failed to rebuild volume backed instance. [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Traceback (most recent call last): [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self.driver.rebuild(**kwargs) [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/driver.py", line 392, in rebuild [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] raise NotImplementedError() [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] NotImplementedError [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] During handling of the above exception, another exception occurred: [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Traceback (most recent call last): [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self._detach_root_volume(context, instance, root_bdm) [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] with excutils.save_and_reraise_exception(): [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self.force_reraise() [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] raise self.value [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self.driver.detach_volume(context, old_connection_info, [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] return self._volumeops.detach_volume(connection_info, instance) [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self._detach_volume_vmdk(connection_info, instance) [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] stable_ref.fetch_moref(session) [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] raise exception.InstanceNotFound(instance_id=self._uuid) [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] nova.exception.InstanceNotFound: Instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 could not be found. [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] During handling of the above exception, another exception occurred: [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Traceback (most recent call last): [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] yield [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 940.503182] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self._do_rebuild_instance_with_claim( [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self._do_rebuild_instance( [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self._rebuild_default_impl(**kwargs) [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] self._rebuild_volume_backed_instance( [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] raise exception.BuildAbortException( [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] nova.exception.BuildAbortException: Build of instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 aborted: Failed to rebuild volume backed instance. [ 940.504459] env[61995]: ERROR nova.compute.manager [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] [ 940.505937] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be751a57-9cf0-4d81-b72d-5399c5743eb3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.532998] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a528b68c-22f8-4a22-a201-6f442a68c7c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.536301] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 940.536301] env[61995]: value = "task-794887" [ 940.536301] env[61995]: _type = "Task" [ 940.536301] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.544238] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210f5abe-9683-43ef-b67e-b56d81853291 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.551160] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794887, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.571510] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa61b67-f45a-4892-b841-7cf614cff993 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.589412] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] The volume has not been displaced from its original location: [datastore2] volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9/volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 940.594804] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Reconfiguring VM instance instance-00000034 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 940.595112] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b66e556-f098-4ebc-beed-0c1e6521460c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.617618] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794882, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.433415} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.619150] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 88638f35-950d-42b4-92b2-ed0dced9148c/88638f35-950d-42b4-92b2-ed0dced9148c.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 940.619396] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.619719] env[61995]: DEBUG oslo_vmware.api [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 940.619719] env[61995]: value = "task-794888" [ 940.619719] env[61995]: _type = "Task" [ 940.619719] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.619920] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-864f1409-4149-4871-9255-6d6af01ec026 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.633166] env[61995]: DEBUG oslo_vmware.api [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794888, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.634603] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 940.634603] env[61995]: value = "task-794889" [ 940.634603] env[61995]: _type = "Task" [ 940.634603] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.649104] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794889, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.765857] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.766477] env[61995]: DEBUG nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 940.769514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.659s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.848035] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.936958] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.937268] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.937436] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.937623] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.937773] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.937926] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.939151] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.939578] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.939684] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.939922] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.940134] env[61995]: DEBUG nova.virt.hardware [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.948346] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 940.948346] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5769157d-af08-42f9-a1f8-1d23e837630f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.974954] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794886, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128435} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.979031] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.979031] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13/d9aa4058-3449-47b3-b10b-42d75d2fcc13.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.979031] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13/d9aa4058-3449-47b3-b10b-42d75d2fcc13.vmdk to [datastore2] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 940.979031] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 940.979031] env[61995]: value = "task-794890" [ 940.979031] env[61995]: _type = "Task" [ 940.979031] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.979031] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8899b174-2012-454a-92e2-f39c1e9be63f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.989954] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794890, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.991751] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 940.991751] env[61995]: value = "task-794891" [ 940.991751] env[61995]: _type = "Task" [ 940.991751] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.001621] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.047667] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794887, 'name': PowerOffVM_Task, 'duration_secs': 0.339504} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.048340] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 941.048340] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 941.048803] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb424aa8-0385-4e95-9277-17056a85753a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.131580] env[61995]: DEBUG oslo_vmware.api [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794888, 'name': ReconfigVM_Task, 'duration_secs': 0.293943} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.131852] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Reconfigured VM instance instance-00000034 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 941.136617] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c99bf2fc-fa87-41f3-8f57-12cac73a1079 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.157367] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794889, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093931} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.157672] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.158575] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb12b96-92cb-4332-83dd-a2259a159f7f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.162803] env[61995]: DEBUG oslo_vmware.api [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 941.162803] env[61995]: value = "task-794893" [ 941.162803] env[61995]: _type = "Task" [ 941.162803] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.185115] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 88638f35-950d-42b4-92b2-ed0dced9148c/88638f35-950d-42b4-92b2-ed0dced9148c.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.186111] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fead6644-676e-4b83-b4c9-ebfddfe2d6a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.203813] env[61995]: DEBUG oslo_vmware.api [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794893, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.211623] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 941.211623] env[61995]: value = "task-794894" [ 941.211623] env[61995]: _type = "Task" [ 941.211623] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.220875] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794894, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.272864] env[61995]: DEBUG nova.compute.utils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 941.283374] env[61995]: DEBUG nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 941.283882] env[61995]: DEBUG nova.network.neutron [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 941.371669] env[61995]: DEBUG nova.policy [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 941.493771] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794890, 'name': ReconfigVM_Task, 'duration_secs': 0.19787} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.497020] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 941.498796] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264acbb2-207f-446f-8c1b-6ddef50ac4d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.521975] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794891, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.534819] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 057ba9d6-4ff2-4da1-be7d-ecfadf738d74/057ba9d6-4ff2-4da1-be7d-ecfadf738d74.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.535509] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04e0f500-4b03-4919-8cd5-269cda4b8f98 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.564277] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 941.564277] env[61995]: value = "task-794895" [ 941.564277] env[61995]: _type = "Task" [ 941.564277] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.576487] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794895, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.677392] env[61995]: DEBUG oslo_vmware.api [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794893, 'name': ReconfigVM_Task, 'duration_secs': 0.190822} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.677749] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185322', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'name': 'volume-34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc7077ac-d3fd-4e84-867a-d86328b32f5d', 'attached_at': '', 'detached_at': '', 'volume_id': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9', 'serial': '34345e8c-85a0-4c01-85e5-bc6ec70b8cc9'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 941.728113] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.783995] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Applying migration context for instance 057ba9d6-4ff2-4da1-be7d-ecfadf738d74 as it has an incoming, in-progress migration f6f8d556-5345-4a35-8742-c2e7db554c68. Migration status is migrating {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 941.786507] env[61995]: INFO nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating resource usage from migration f6f8d556-5345-4a35-8742-c2e7db554c68 [ 941.789521] env[61995]: DEBUG nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 941.814870] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d9c53806-68c0-4872-a262-36bc05573674 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.815117] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance dc7077ac-d3fd-4e84-867a-d86328b32f5d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.815254] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b40b1866-ced3-40ef-9ab7-30d58b75f288 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.815789] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 4edbf65d-eaca-47e1-b6bf-03ccb908b52e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.816062] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 229d966c-b70d-4a51-a176-2e88488d5c4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.816210] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ed952baf-f1ef-4ec0-8a15-928287190456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.816298] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e835ac53-29af-4bd0-b186-5c6270ccf760 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.816474] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 5a6bd8a7-2633-42d2-b853-4e3a5690c77e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.816597] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 7d987006-afac-4973-ae69-f15b4752469b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.816713] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.816968] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.817111] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 88638f35-950d-42b4-92b2-ed0dced9148c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.817229] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Migration f6f8d556-5345-4a35-8742-c2e7db554c68 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 941.817341] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 057ba9d6-4ff2-4da1-be7d-ecfadf738d74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.817451] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 453f46c7-5bad-4ca2-b228-f76e62fbd03e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 941.817681] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 941.817864] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3712MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 941.899947] env[61995]: DEBUG nova.network.neutron [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Successfully created port: cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.926214] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.926567] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.926818] env[61995]: INFO nova.compute.manager [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Attaching volume 908f434d-3808-42d4-a891-104337fb5452 to /dev/sdb [ 941.976284] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5304683-0104-4e88-978b-ccf5c1e0790e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.991027] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d486185-d4ba-44be-a568-69dae64d7c76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.009212] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794891, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.017628] env[61995]: DEBUG nova.virt.block_device [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating existing volume attachment record: 1ee3f09e-0588-4e7e-82d0-d00302ea20f0 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 942.051957] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 942.052287] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 942.052523] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Deleting the datastore file [datastore2] 4edbf65d-eaca-47e1-b6bf-03ccb908b52e {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.052807] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e82ad7cf-5d20-463e-a443-1419232e65b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.062722] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for the task: (returnval){ [ 942.062722] env[61995]: value = "task-794896" [ 942.062722] env[61995]: _type = "Task" [ 942.062722] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.079666] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.083513] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794895, 'name': ReconfigVM_Task, 'duration_secs': 0.449244} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.083915] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 057ba9d6-4ff2-4da1-be7d-ecfadf738d74/057ba9d6-4ff2-4da1-be7d-ecfadf738d74.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.084244] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance '057ba9d6-4ff2-4da1-be7d-ecfadf738d74' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 942.144150] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20329c0e-a91d-4e21-b888-406064dc5dc6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.158557] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e421327c-2495-4afc-a919-ba89cd5c79db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.199613] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3edbd51c-28a3-438b-b47f-2cc7f4c59c1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.213867] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa69355-3fd4-4bf2-94f1-423447e769e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.232926] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.244785] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.248341] env[61995]: DEBUG nova.objects.instance [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'flavor' on Instance uuid dc7077ac-d3fd-4e84-867a-d86328b32f5d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.510884] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794891, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.541862] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.575930] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.598728] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93eee522-842a-46bc-9982-03f497bf471a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.624539] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d80b70-d7f7-4a08-9a64-462cfed0e782 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.647214] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance '057ba9d6-4ff2-4da1-be7d-ecfadf738d74' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 942.734777] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.748229] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.805041] env[61995]: DEBUG nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 942.834759] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.835157] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.835229] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.835401] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.835592] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.835753] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.835992] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.836285] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.836470] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.836640] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.836841] env[61995]: DEBUG nova.virt.hardware [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.837764] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938db8c3-be24-44c3-b7b1-a25f29ec5fcd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.849862] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b9202f-c5cf-492d-ab7a-edba162b7177 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.009929] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794891, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.081375] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.234393] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.254582] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 943.254846] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.485s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.255250] env[61995]: DEBUG oslo_concurrency.lockutils [None req-18c5f8ed-6be4-4ed5-888c-e896ec9f9cda tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.320s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.256452] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.715s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.264887] env[61995]: DEBUG nova.network.neutron [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Port 0496010d-b03c-4002-b2d4-372cdbededa9 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 943.510160] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794891, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.529366] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c6c01a-c864-4db2-b467-347fefd82238 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.542676] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30825a4a-a26a-451a-8e2b-7ef740268579 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.580077] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.580375] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.580595] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.580784] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.580957] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.583214] env[61995]: INFO nova.compute.manager [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Terminating instance [ 943.590142] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92e2261-002d-496f-8b86-b91c00761c99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.594387] env[61995]: DEBUG nova.compute.manager [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 943.594614] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 943.595439] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af858c00-b1f6-4768-a86a-b31443795feb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.610266] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3f31fd-4a79-49f3-9bdf-9c366df2ff60 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.614602] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.617615] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 943.618545] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b23cf4a3-f268-46b3-b51a-71bf87449223 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.635714] env[61995]: DEBUG nova.compute.provider_tree [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.639759] env[61995]: DEBUG oslo_vmware.api [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 943.639759] env[61995]: value = "task-794900" [ 943.639759] env[61995]: _type = "Task" [ 943.639759] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.654553] env[61995]: DEBUG oslo_vmware.api [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.683011] env[61995]: DEBUG nova.compute.manager [req-e3f080c8-cc7e-44e5-8eaf-fac81c250b85 req-3d3b56fa-c857-4618-bd0b-c306ff29f1d2 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-vif-plugged-cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.683359] env[61995]: DEBUG oslo_concurrency.lockutils [req-e3f080c8-cc7e-44e5-8eaf-fac81c250b85 req-3d3b56fa-c857-4618-bd0b-c306ff29f1d2 service nova] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.683631] env[61995]: DEBUG oslo_concurrency.lockutils [req-e3f080c8-cc7e-44e5-8eaf-fac81c250b85 req-3d3b56fa-c857-4618-bd0b-c306ff29f1d2 service nova] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.683780] env[61995]: DEBUG oslo_concurrency.lockutils [req-e3f080c8-cc7e-44e5-8eaf-fac81c250b85 req-3d3b56fa-c857-4618-bd0b-c306ff29f1d2 service nova] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.683992] env[61995]: DEBUG nova.compute.manager [req-e3f080c8-cc7e-44e5-8eaf-fac81c250b85 req-3d3b56fa-c857-4618-bd0b-c306ff29f1d2 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] No waiting events found dispatching network-vif-plugged-cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.684467] env[61995]: WARNING nova.compute.manager [req-e3f080c8-cc7e-44e5-8eaf-fac81c250b85 req-3d3b56fa-c857-4618-bd0b-c306ff29f1d2 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received unexpected event network-vif-plugged-cdaef75a-1442-4f8d-8eab-340da21a74d5 for instance with vm_state building and task_state spawning. [ 943.734653] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794894, 'name': ReconfigVM_Task, 'duration_secs': 2.25041} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.735079] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 88638f35-950d-42b4-92b2-ed0dced9148c/88638f35-950d-42b4-92b2-ed0dced9148c.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.736183] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e0dd01d-6c64-437a-a006-5e7bcdf24466 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.743754] env[61995]: DEBUG nova.network.neutron [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Successfully updated port: cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.750586] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 943.750586] env[61995]: value = "task-794901" [ 943.750586] env[61995]: _type = "Task" [ 943.750586] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.765696] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794901, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.968197] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquiring lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.968660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.969117] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquiring lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.969377] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.969584] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.974704] env[61995]: INFO nova.compute.manager [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Terminating instance [ 943.976961] env[61995]: DEBUG nova.compute.manager [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 943.977357] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92e59b85-fbe7-44fe-aa97-24da432067d9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.991524] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafd7032-e5d4-49db-8606-bafc6f0af0bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.020748] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794891, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.878085} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.021196] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d9aa4058-3449-47b3-b10b-42d75d2fcc13/d9aa4058-3449-47b3-b10b-42d75d2fcc13.vmdk to [datastore2] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 944.022038] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404b2501-5cce-4edc-a884-fdfbdb37eac7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.047322] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.057921] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa002623-847c-4a03-bf1a-01e4771c404d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.073923] env[61995]: WARNING nova.virt.vmwareapi.driver [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 could not be found. [ 944.074200] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 944.074611] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78d681b1-b04c-46da-b161-0a68439964d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.089128] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 944.089128] env[61995]: value = "task-794902" [ 944.089128] env[61995]: _type = "Task" [ 944.089128] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.092827] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb15080-7222-4227-8427-ef96deb547a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.111908] env[61995]: DEBUG oslo_vmware.api [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Task: {'id': task-794896, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.815079} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.112792] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.113062] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 944.113351] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 944.113495] env[61995]: INFO nova.compute.manager [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Took 3.63 seconds to destroy the instance on the hypervisor. [ 944.113804] env[61995]: DEBUG oslo.service.loopingcall [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.114429] env[61995]: DEBUG nova.compute.manager [-] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 944.114583] env[61995]: DEBUG nova.network.neutron [-] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 944.120162] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794902, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.136617] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 could not be found. [ 944.136978] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 944.137158] env[61995]: INFO nova.compute.manager [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Took 0.16 seconds to destroy the instance on the hypervisor. [ 944.137361] env[61995]: DEBUG oslo.service.loopingcall [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.137652] env[61995]: DEBUG nova.compute.manager [-] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 944.137757] env[61995]: DEBUG nova.network.neutron [-] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 944.141578] env[61995]: DEBUG nova.scheduler.client.report [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.154129] env[61995]: DEBUG oslo_vmware.api [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794900, 'name': PowerOffVM_Task, 'duration_secs': 0.418085} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.154465] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 944.154668] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 944.155057] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-675eaacb-711b-488a-9ee1-8ebc01395d9f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.236284] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 944.236701] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 944.236996] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleting the datastore file [datastore2] dc7077ac-d3fd-4e84-867a-d86328b32f5d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.237395] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab1c86ac-2da8-46f8-a651-e6c60bb60ee9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.247121] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.247301] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.247421] env[61995]: DEBUG nova.network.neutron [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.248990] env[61995]: DEBUG oslo_vmware.api [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 944.248990] env[61995]: value = "task-794904" [ 944.248990] env[61995]: _type = "Task" [ 944.248990] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.265126] env[61995]: DEBUG oslo_vmware.api [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.265466] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794901, 'name': Rename_Task, 'duration_secs': 0.288942} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.265949] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 944.266628] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03316ad5-2378-4b6f-8131-616c2876e57b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.283827] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 944.283827] env[61995]: value = "task-794905" [ 944.283827] env[61995]: _type = "Task" [ 944.283827] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.291724] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.291978] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.292294] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.305768] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794905, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.582021] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ff665d-f6a7-d542-4a5e-a8cbe9f8c9c4/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 944.582021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38bd9440-8e8d-4d77-9a50-44c756d2cc64 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.588328] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ff665d-f6a7-d542-4a5e-a8cbe9f8c9c4/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 944.588618] env[61995]: ERROR oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ff665d-f6a7-d542-4a5e-a8cbe9f8c9c4/disk-0.vmdk due to incomplete transfer. [ 944.588884] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-34c66ce9-743f-46df-aadb-387dee227d86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.598742] env[61995]: DEBUG oslo_vmware.rw_handles [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ff665d-f6a7-d542-4a5e-a8cbe9f8c9c4/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 944.599028] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Uploaded image 91a96abd-c3b8-4135-bd90-9f82c94fdfcb to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 944.602993] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 944.602993] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-63d25704-b148-43ea-9be4-8c142f6c7dbd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.619744] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794902, 'name': ReconfigVM_Task, 'duration_secs': 0.408786} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.624671] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5/0e152049-f9ca-4ac1-b12d-ae1ad4623ff5.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.624671] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 944.624671] env[61995]: value = "task-794907" [ 944.624671] env[61995]: _type = "Task" [ 944.624671] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.624671] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29afb259-3934-4896-ab9a-b1f98537fa74 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.637050] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794907, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.638786] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 944.638786] env[61995]: value = "task-794908" [ 944.638786] env[61995]: _type = "Task" [ 944.638786] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.650339] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.394s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.650608] env[61995]: INFO nova.compute.manager [None req-0963eb2f-e190-4aa2-ab41-72a9eaadd5e4 tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Successfully reverted task state from rebuilding on failure for instance. [ 944.657225] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794908, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.771823] env[61995]: DEBUG oslo_vmware.api [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238868} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.772391] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.772762] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 944.773136] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 944.777034] env[61995]: INFO nova.compute.manager [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Took 1.18 seconds to destroy the instance on the hypervisor. [ 944.777034] env[61995]: DEBUG oslo.service.loopingcall [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.777034] env[61995]: DEBUG nova.compute.manager [-] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 944.777034] env[61995]: DEBUG nova.network.neutron [-] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 944.800125] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794905, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.802534] env[61995]: DEBUG nova.network.neutron [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 944.905145] env[61995]: DEBUG nova.network.neutron [-] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.034091] env[61995]: DEBUG nova.network.neutron [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.137136] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794907, 'name': Destroy_Task, 'duration_secs': 0.445831} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.138210] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Destroyed the VM [ 945.138210] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 945.138431] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0fbc9021-c0e2-4b36-a10d-008dee11d3bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.152257] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794908, 'name': Rename_Task, 'duration_secs': 0.440732} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.153776] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 945.156347] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 945.156347] env[61995]: value = "task-794909" [ 945.156347] env[61995]: _type = "Task" [ 945.156347] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.156347] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a495048-1c94-43e3-a018-1fea1ae1a67b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.167927] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794909, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.170254] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 945.170254] env[61995]: value = "task-794910" [ 945.170254] env[61995]: _type = "Task" [ 945.170254] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.180103] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794910, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.299048] env[61995]: DEBUG oslo_vmware.api [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794905, 'name': PowerOnVM_Task, 'duration_secs': 0.778014} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.302204] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 945.302204] env[61995]: INFO nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Took 10.67 seconds to spawn the instance on the hypervisor. [ 945.302204] env[61995]: DEBUG nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.305869] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f52aa4b-f73a-4797-a7a1-b0b90b5063d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.353697] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.354646] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.354646] env[61995]: DEBUG nova.network.neutron [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 945.408923] env[61995]: INFO nova.compute.manager [-] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Took 1.29 seconds to deallocate network for instance. [ 945.441424] env[61995]: DEBUG nova.compute.manager [req-7cabfdf2-0174-43e7-83f0-24a38ff9f6f1 req-ac3ed1ee-4eff-4327-8640-293a79e5f888 service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Received event network-vif-deleted-ea813bd4-7083-4f67-b0ca-89282c951487 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.441642] env[61995]: INFO nova.compute.manager [req-7cabfdf2-0174-43e7-83f0-24a38ff9f6f1 req-ac3ed1ee-4eff-4327-8640-293a79e5f888 service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Neutron deleted interface ea813bd4-7083-4f67-b0ca-89282c951487; detaching it from the instance and deleting it from the info cache [ 945.441805] env[61995]: DEBUG nova.network.neutron [req-7cabfdf2-0174-43e7-83f0-24a38ff9f6f1 req-ac3ed1ee-4eff-4327-8640-293a79e5f888 service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.536333] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.539023] env[61995]: DEBUG nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Instance network_info: |[{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.539023] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:10:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdaef75a-1442-4f8d-8eab-340da21a74d5', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.548845] env[61995]: DEBUG oslo.service.loopingcall [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.549381] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 945.549653] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dbe232ca-f45b-4dae-9ef9-0972e0b7851a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.582619] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.582619] env[61995]: value = "task-794911" [ 945.582619] env[61995]: _type = "Task" [ 945.582619] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.590554] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794911, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.672522] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794909, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.686925] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794910, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.836515] env[61995]: INFO nova.compute.manager [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Took 20.03 seconds to build instance. [ 945.888586] env[61995]: DEBUG nova.compute.manager [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-changed-cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.888586] env[61995]: DEBUG nova.compute.manager [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing instance network info cache due to event network-changed-cdaef75a-1442-4f8d-8eab-340da21a74d5. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 945.888586] env[61995]: DEBUG oslo_concurrency.lockutils [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.888586] env[61995]: DEBUG oslo_concurrency.lockutils [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.888586] env[61995]: DEBUG nova.network.neutron [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing network info cache for port cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.901400] env[61995]: DEBUG nova.network.neutron [-] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.922218] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.922610] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.922994] env[61995]: DEBUG nova.objects.instance [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lazy-loading 'resources' on Instance uuid 4edbf65d-eaca-47e1-b6bf-03ccb908b52e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.948131] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9350a93-fbb8-474c-a0c4-2de29b475b0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.959534] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f1186d-4a92-4c6f-852f-11a51d0481cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.001258] env[61995]: DEBUG nova.compute.manager [req-7cabfdf2-0174-43e7-83f0-24a38ff9f6f1 req-ac3ed1ee-4eff-4327-8640-293a79e5f888 service nova] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Detach interface failed, port_id=ea813bd4-7083-4f67-b0ca-89282c951487, reason: Instance 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 946.091481] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794911, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.171848] env[61995]: DEBUG oslo_vmware.api [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794909, 'name': RemoveSnapshot_Task, 'duration_secs': 0.562733} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.171848] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 946.172044] env[61995]: INFO nova.compute.manager [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Took 18.82 seconds to snapshot the instance on the hypervisor. [ 946.190378] env[61995]: DEBUG oslo_vmware.api [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794910, 'name': PowerOnVM_Task, 'duration_secs': 0.812087} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.190669] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 946.322272] env[61995]: DEBUG nova.compute.manager [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.323273] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a002794-f2c4-4a72-b739-10d7de3a510d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.327590] env[61995]: DEBUG nova.network.neutron [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance_info_cache with network_info: [{"id": "0496010d-b03c-4002-b2d4-372cdbededa9", "address": "fa:16:3e:6d:e6:cd", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0496010d-b0", "ovs_interfaceid": "0496010d-b03c-4002-b2d4-372cdbededa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.344960] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c9e4ab51-3790-4b2d-afae-610a4c49d9e5 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "88638f35-950d-42b4-92b2-ed0dced9148c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.551s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.403619] env[61995]: INFO nova.compute.manager [-] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Took 2.27 seconds to deallocate network for instance. [ 946.597938] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794911, 'name': CreateVM_Task, 'duration_secs': 0.561286} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.602829] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 946.605369] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.605552] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.605961] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.606665] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-deb5cc2e-da17-42f7-8969-a05004849265 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.612682] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 946.612682] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f57171-4512-b997-bd27-f219e2c281a3" [ 946.612682] env[61995]: _type = "Task" [ 946.612682] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.622018] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f57171-4512-b997-bd27-f219e2c281a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.627599] env[61995]: DEBUG nova.network.neutron [-] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.689059] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8521a79b-3318-4bf0-a906-ba307fd1e4b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.702083] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df255d4-e1a4-4bd5-a348-4a3325036083 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.741582] env[61995]: DEBUG nova.network.neutron [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updated VIF entry in instance network info cache for port cdaef75a-1442-4f8d-8eab-340da21a74d5. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 946.742356] env[61995]: DEBUG nova.network.neutron [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.744434] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a160ae-2fc8-4a21-b7cf-233c4dfd2574 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.753531] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df56b0c5-8d49-4b52-bbe7-4d68cf79c000 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.762722] env[61995]: DEBUG nova.compute.manager [None req-a29c4c8b-9c5a-426b-9453-f58a1f2b2e49 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Found 2 images (rotation: 2) {{(pid=61995) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 946.776954] env[61995]: DEBUG nova.compute.provider_tree [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.832940] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.848765] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b188594c-bb31-4f91-a6c3-8573233340f9 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 30.524s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.971694] env[61995]: INFO nova.compute.manager [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Took 0.57 seconds to detach 1 volumes for instance. [ 946.974737] env[61995]: DEBUG nova.compute.manager [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Deleting volume: 604ce120-96fa-4557-9558-055ca5966123 {{(pid=61995) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 947.094613] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 947.095143] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185367', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'name': 'volume-908f434d-3808-42d4-a891-104337fb5452', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e835ac53-29af-4bd0-b186-5c6270ccf760', 'attached_at': '', 'detached_at': '', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'serial': '908f434d-3808-42d4-a891-104337fb5452'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 947.096583] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b976daa7-76d3-4ac9-af8b-133b722d2604 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.119721] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7602e605-43d9-4c7d-b1e4-9c6927c5b584 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.129876] env[61995]: INFO nova.compute.manager [-] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Took 2.36 seconds to deallocate network for instance. [ 947.130308] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f57171-4512-b997-bd27-f219e2c281a3, 'name': SearchDatastore_Task, 'duration_secs': 0.029967} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.146029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.146029] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.146029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.146029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.146259] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.154242] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] volume-908f434d-3808-42d4-a891-104337fb5452/volume-908f434d-3808-42d4-a891-104337fb5452.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.157530] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3416a120-b8eb-4681-8902-bc997bda01c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.158875] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc4034d7-7c0c-4a14-af71-56b18733714e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.187757] env[61995]: DEBUG oslo_vmware.api [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 947.187757] env[61995]: value = "task-794913" [ 947.187757] env[61995]: _type = "Task" [ 947.187757] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.191817] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.194941] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 947.194941] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aed11604-b0b9-45e5-ad16-ae12bd92cae2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.201438] env[61995]: DEBUG oslo_vmware.api [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.207472] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 947.207472] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52110cab-585e-df0d-cfb3-7681ac4ccbce" [ 947.207472] env[61995]: _type = "Task" [ 947.207472] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.221924] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52110cab-585e-df0d-cfb3-7681ac4ccbce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.248102] env[61995]: DEBUG oslo_concurrency.lockutils [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.248469] env[61995]: DEBUG nova.compute.manager [req-ab6d9721-ae01-4ea4-bb5b-a67f7ef1d845 req-8336b936-4937-4d09-bfb5-a8fe2b6fd063 service nova] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Received event network-vif-deleted-591f35b8-baf5-4da1-b479-dc99ae2f7b75 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.281343] env[61995]: DEBUG nova.scheduler.client.report [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.391854] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f067611-9c01-4261-9f5e-2f0df64dcd5e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.418033] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082ec9ac-ba41-4647-a116-056d2fb33dbd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.426805] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance '057ba9d6-4ff2-4da1-be7d-ecfadf738d74' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 947.468811] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "24b974c0-9567-46a3-96f6-fac1e602f46e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.469177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.546219] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.569189] env[61995]: DEBUG nova.compute.manager [req-10462fa8-c8cf-43d8-94bc-3cc76ee5e56b req-b02fe47d-5c7f-45a4-9da8-85d002c5f1d8 service nova] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Received event network-vif-deleted-b970a5e5-c432-4698-bc92-c64f8b7e3091 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.588757] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "88638f35-950d-42b4-92b2-ed0dced9148c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.589058] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "88638f35-950d-42b4-92b2-ed0dced9148c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.589321] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "88638f35-950d-42b4-92b2-ed0dced9148c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.589939] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "88638f35-950d-42b4-92b2-ed0dced9148c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.590180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "88638f35-950d-42b4-92b2-ed0dced9148c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.592433] env[61995]: INFO nova.compute.manager [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Terminating instance [ 947.598935] env[61995]: DEBUG nova.compute.manager [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.599559] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.600669] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523422b2-2747-463a-a48d-ca086b26f0e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.609783] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.609783] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a8eca4a-8391-49ce-8fc8-e90e3302f84e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.622931] env[61995]: DEBUG oslo_vmware.api [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 947.622931] env[61995]: value = "task-794914" [ 947.622931] env[61995]: _type = "Task" [ 947.622931] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.636967] env[61995]: DEBUG oslo_vmware.api [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.681188] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.701636] env[61995]: DEBUG oslo_vmware.api [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794913, 'name': ReconfigVM_Task, 'duration_secs': 0.383486} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.705837] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Reconfigured VM instance instance-00000047 to attach disk [datastore1] volume-908f434d-3808-42d4-a891-104337fb5452/volume-908f434d-3808-42d4-a891-104337fb5452.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.722161] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f70d3b9e-1d21-4b08-a248-3ac178727198 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.745681] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52110cab-585e-df0d-cfb3-7681ac4ccbce, 'name': SearchDatastore_Task, 'duration_secs': 0.015434} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.748318] env[61995]: DEBUG oslo_vmware.api [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 947.748318] env[61995]: value = "task-794915" [ 947.748318] env[61995]: _type = "Task" [ 947.748318] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.748649] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87db2c83-14ec-4d28-9aff-f3ffd112b497 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.761689] env[61995]: DEBUG oslo_vmware.api [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.763321] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 947.763321] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c3edf0-cf1c-24d0-8956-7ae5f97fb953" [ 947.763321] env[61995]: _type = "Task" [ 947.763321] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.772816] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c3edf0-cf1c-24d0-8956-7ae5f97fb953, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.784953] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.788748] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.242s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.788748] env[61995]: DEBUG nova.objects.instance [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lazy-loading 'resources' on Instance uuid 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.819703] env[61995]: INFO nova.scheduler.client.report [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Deleted allocations for instance 4edbf65d-eaca-47e1-b6bf-03ccb908b52e [ 947.935528] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.936193] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7420c915-843f-4f4f-b03e-b98b42eb8d29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.946302] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 947.946302] env[61995]: value = "task-794916" [ 947.946302] env[61995]: _type = "Task" [ 947.946302] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.957956] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.972086] env[61995]: DEBUG nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.143691] env[61995]: DEBUG oslo_vmware.api [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794914, 'name': PowerOffVM_Task, 'duration_secs': 0.331627} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.147508] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 948.148060] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 948.150390] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.150640] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.152806] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc892a02-28e3-490e-b360-8ba28a1adfe7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.258684] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 948.259437] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 948.259757] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleting the datastore file [datastore2] 88638f35-950d-42b4-92b2-ed0dced9148c {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.260569] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78bfd9df-9a42-4890-9018-3c32c5703263 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.266816] env[61995]: DEBUG oslo_vmware.api [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794915, 'name': ReconfigVM_Task, 'duration_secs': 0.317538} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.270560] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185367', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'name': 'volume-908f434d-3808-42d4-a891-104337fb5452', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e835ac53-29af-4bd0-b186-5c6270ccf760', 'attached_at': '', 'detached_at': '', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'serial': '908f434d-3808-42d4-a891-104337fb5452'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 948.276026] env[61995]: DEBUG oslo_vmware.api [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 948.276026] env[61995]: value = "task-794918" [ 948.276026] env[61995]: _type = "Task" [ 948.276026] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.279276] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c3edf0-cf1c-24d0-8956-7ae5f97fb953, 'name': SearchDatastore_Task, 'duration_secs': 0.017155} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.283025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.283025] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 453f46c7-5bad-4ca2-b228-f76e62fbd03e/453f46c7-5bad-4ca2-b228-f76e62fbd03e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 948.283256] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-051b7501-8ea7-4da1-862f-9e19f02b880a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.294985] env[61995]: DEBUG oslo_vmware.api [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794918, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.296769] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 948.296769] env[61995]: value = "task-794919" [ 948.296769] env[61995]: _type = "Task" [ 948.296769] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.308148] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794919, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.332802] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d3bb1b2a-2e5c-4761-9408-b1c53311cf59 tempest-MigrationsAdminTest-1313296370 tempest-MigrationsAdminTest-1313296370-project-member] Lock "4edbf65d-eaca-47e1-b6bf-03ccb908b52e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.858s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.466289] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794916, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.494791] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.565885] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675cbd70-f45a-428d-9034-a582f4d90f1b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.574813] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fb7c33-b277-46be-8bcd-6076b235ea90 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.610923] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982b11a6-d6f3-422c-83b6-13c2bafec88b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.621290] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4b1763-3ef9-4659-952d-606be00640ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.636464] env[61995]: DEBUG nova.compute.manager [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.637053] env[61995]: DEBUG nova.compute.provider_tree [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.643495] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c615135-f427-4e94-91ac-d348c6df5085 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.655032] env[61995]: DEBUG nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.792795] env[61995]: DEBUG oslo_vmware.api [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794918, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197218} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.793190] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.793511] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 948.793779] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 948.794022] env[61995]: INFO nova.compute.manager [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Took 1.19 seconds to destroy the instance on the hypervisor. [ 948.794372] env[61995]: DEBUG oslo.service.loopingcall [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.794673] env[61995]: DEBUG nova.compute.manager [-] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.794838] env[61995]: DEBUG nova.network.neutron [-] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.814027] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794919, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.870891] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.871277] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.871542] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.871789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.871995] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.874750] env[61995]: INFO nova.compute.manager [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Terminating instance [ 948.878408] env[61995]: DEBUG nova.compute.manager [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 948.878659] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 948.879732] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ab4fe9-e48d-4c11-bf3a-6bd1dbcc6a5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.895070] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 948.897526] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47adc443-c20e-4d05-bbb3-fc5f8ad4a2d1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.907029] env[61995]: DEBUG oslo_vmware.api [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 948.907029] env[61995]: value = "task-794920" [ 948.907029] env[61995]: _type = "Task" [ 948.907029] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.919831] env[61995]: DEBUG oslo_vmware.api [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.971602] env[61995]: DEBUG oslo_vmware.api [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794916, 'name': PowerOnVM_Task, 'duration_secs': 0.56748} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.971602] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 948.971602] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cb14441a-8e07-4f60-9bd2-bd3c6d22c162 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance '057ba9d6-4ff2-4da1-be7d-ecfadf738d74' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 949.144428] env[61995]: DEBUG nova.scheduler.client.report [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.166588] env[61995]: INFO nova.compute.manager [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] instance snapshotting [ 949.167555] env[61995]: DEBUG nova.objects.instance [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'flavor' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.194385] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.309636] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794919, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627171} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.309928] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 453f46c7-5bad-4ca2-b228-f76e62fbd03e/453f46c7-5bad-4ca2-b228-f76e62fbd03e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 949.310164] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.310446] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-896bb04d-39de-49a1-86fc-f961f218bf99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.317862] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 949.317862] env[61995]: value = "task-794921" [ 949.317862] env[61995]: _type = "Task" [ 949.317862] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.327017] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794921, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.330046] env[61995]: DEBUG nova.objects.instance [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.416358] env[61995]: DEBUG oslo_vmware.api [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794920, 'name': PowerOffVM_Task, 'duration_secs': 0.280738} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.416640] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.416812] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.417138] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3848f602-a7d9-4c37-afb2-1f00b27fe881 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.499156] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.499397] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.499612] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Deleting the datastore file [datastore1] 5a6bd8a7-2633-42d2-b853-4e3a5690c77e {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.499868] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00d4e453-84bd-4b76-8032-f2fe4b226a15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.508628] env[61995]: DEBUG oslo_vmware.api [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for the task: (returnval){ [ 949.508628] env[61995]: value = "task-794923" [ 949.508628] env[61995]: _type = "Task" [ 949.508628] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.517971] env[61995]: DEBUG oslo_vmware.api [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.625809] env[61995]: DEBUG nova.network.neutron [-] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.655022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.864s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.655022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.973s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.655022] env[61995]: DEBUG nova.objects.instance [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'resources' on Instance uuid dc7077ac-d3fd-4e84-867a-d86328b32f5d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.664163] env[61995]: DEBUG nova.compute.manager [req-e63fd05c-250f-4c77-b824-a760e7a80938 req-ba48eb0a-76a8-455e-8c38-5b101c3256a6 service nova] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Received event network-vif-deleted-272691df-c5ce-4116-a3e1-5e15186723b5 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.677480] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbcab26a-16ce-4448-b9e0-dc58de9832b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.702018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f64a0e1-d1eb-4d7d-9ed4-2a927330949f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.737432] env[61995]: DEBUG oslo_concurrency.lockutils [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.829248] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794921, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074822} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.829248] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.831126] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d22b721-f06b-4805-bc83-2e2e2d044bcc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.835102] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c114ac57-c2c1-46b5-85b5-6cb89e3ad7bb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.909s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.836095] env[61995]: DEBUG oslo_concurrency.lockutils [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.099s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.836341] env[61995]: DEBUG nova.compute.manager [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.837409] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e160b8b4-c01d-4e14-a438-1c04c2b18b8b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.857680] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 453f46c7-5bad-4ca2-b228-f76e62fbd03e/453f46c7-5bad-4ca2-b228-f76e62fbd03e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.859081] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa948148-207e-495f-8c3f-abc83867fee8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.876590] env[61995]: DEBUG nova.compute.manager [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 949.876663] env[61995]: DEBUG nova.objects.instance [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.884876] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 949.884876] env[61995]: value = "task-794924" [ 949.884876] env[61995]: _type = "Task" [ 949.884876] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.894917] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794924, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.005950] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a216871-aead-4134-932f-1c05be982c7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.017670] env[61995]: DEBUG oslo_vmware.api [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.019934] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8581ae-d7c1-430a-83f5-5d03f352c9e2 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Suspending the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 950.020191] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-c5b85460-12f8-4768-83bb-5b5df1851dee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.027643] env[61995]: DEBUG oslo_vmware.api [None req-5b8581ae-d7c1-430a-83f5-5d03f352c9e2 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 950.027643] env[61995]: value = "task-794925" [ 950.027643] env[61995]: _type = "Task" [ 950.027643] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.038353] env[61995]: DEBUG oslo_vmware.api [None req-5b8581ae-d7c1-430a-83f5-5d03f352c9e2 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794925, 'name': SuspendVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.128602] env[61995]: INFO nova.compute.manager [-] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Took 1.33 seconds to deallocate network for instance. [ 950.175074] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91816f99-f284-41bc-a756-7a68aa85146c tempest-ServerActionsV293TestJSON-626651582 tempest-ServerActionsV293TestJSON-626651582-project-member] Lock "4ca0dcd7-a2b8-4de6-835b-07c36ab95a99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.206s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.213336] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 950.213779] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ac3f5eeb-dba9-4fc4-b317-c466ecc7ebb1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.223028] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 950.223028] env[61995]: value = "task-794926" [ 950.223028] env[61995]: _type = "Task" [ 950.223028] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.236675] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794926, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.382331] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.382710] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33645c7f-b797-4dac-8c5c-94d7eaa31d54 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.395016] env[61995]: DEBUG oslo_vmware.api [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 950.395016] env[61995]: value = "task-794927" [ 950.395016] env[61995]: _type = "Task" [ 950.395016] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.403604] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794924, 'name': ReconfigVM_Task, 'duration_secs': 0.33301} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.405999] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 453f46c7-5bad-4ca2-b228-f76e62fbd03e/453f46c7-5bad-4ca2-b228-f76e62fbd03e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.407250] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1016e52f-0af2-40cc-992e-ca73aa0e9f66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.412946] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d68e4f83-40c1-4fa9-a335-2eb854324655 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.414687] env[61995]: DEBUG oslo_vmware.api [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.420816] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229db649-c4c8-489d-ba8a-86cfaee92d81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.425922] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 950.425922] env[61995]: value = "task-794928" [ 950.425922] env[61995]: _type = "Task" [ 950.425922] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.460217] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e946ad5-1097-433d-a0ca-d127cb70c0a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.464471] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794928, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.471207] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1656cd-b1f4-4ca4-b86c-8c0d64ffe33b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.491199] env[61995]: DEBUG nova.compute.provider_tree [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.523471] env[61995]: DEBUG oslo_vmware.api [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Task: {'id': task-794923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.594865} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.523779] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.523982] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.524184] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.524386] env[61995]: INFO nova.compute.manager [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Took 1.65 seconds to destroy the instance on the hypervisor. [ 950.524668] env[61995]: DEBUG oslo.service.loopingcall [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.525400] env[61995]: DEBUG nova.compute.manager [-] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.525507] env[61995]: DEBUG nova.network.neutron [-] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.538164] env[61995]: DEBUG oslo_vmware.api [None req-5b8581ae-d7c1-430a-83f5-5d03f352c9e2 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794925, 'name': SuspendVM_Task} progress is 58%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.637380] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.739026] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794926, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.906275] env[61995]: DEBUG oslo_vmware.api [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794927, 'name': PowerOffVM_Task, 'duration_secs': 0.261936} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.906560] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.906748] env[61995]: DEBUG nova.compute.manager [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.907953] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdef0a0-36d9-404d-b346-96239afd0126 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.936503] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794928, 'name': Rename_Task, 'duration_secs': 0.156604} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.936805] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.937618] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-413bd257-304a-4493-8299-44cf5f301fb7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.944803] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 950.944803] env[61995]: value = "task-794929" [ 950.944803] env[61995]: _type = "Task" [ 950.944803] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.953972] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794929, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.995655] env[61995]: DEBUG nova.scheduler.client.report [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.038316] env[61995]: DEBUG oslo_vmware.api [None req-5b8581ae-d7c1-430a-83f5-5d03f352c9e2 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794925, 'name': SuspendVM_Task, 'duration_secs': 0.784434} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.038610] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8581ae-d7c1-430a-83f5-5d03f352c9e2 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Suspended the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 951.038903] env[61995]: DEBUG nova.compute.manager [None req-5b8581ae-d7c1-430a-83f5-5d03f352c9e2 tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.039579] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e14d08-9846-44fe-9c58-ba11adc2b364 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.238605] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794926, 'name': CreateSnapshot_Task, 'duration_secs': 0.586425} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.241092] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 951.241092] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2314c9db-3b57-4b02-890c-795abb44954a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.421425] env[61995]: DEBUG oslo_concurrency.lockutils [None req-acf67892-539b-44d7-b7f2-f3be0e5951ac tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.584s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.460935] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794929, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.501080] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.847s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.504132] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.010s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.506074] env[61995]: INFO nova.compute.claims [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.530706] env[61995]: INFO nova.scheduler.client.report [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted allocations for instance dc7077ac-d3fd-4e84-867a-d86328b32f5d [ 951.593022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.593428] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.593508] env[61995]: DEBUG nova.compute.manager [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Going to confirm migration 4 {{(pid=61995) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 951.647375] env[61995]: DEBUG nova.network.neutron [-] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.695808] env[61995]: DEBUG nova.compute.manager [req-c620704a-e509-4cd0-8bac-3ca2f2fab353 req-306ff008-e887-40ac-8526-65c3e1823f5b service nova] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Received event network-vif-deleted-05b0b88f-30eb-4e40-bf1d-5ee6cc2ea82f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.761114] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 951.761621] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-06971ee2-61f7-4b03-9a03-5d3ce07ae316 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.772366] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 951.772366] env[61995]: value = "task-794930" [ 951.772366] env[61995]: _type = "Task" [ 951.772366] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.786603] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794930, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.962342] env[61995]: DEBUG oslo_vmware.api [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-794929, 'name': PowerOnVM_Task, 'duration_secs': 0.538776} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.962578] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 951.962794] env[61995]: INFO nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Took 9.16 seconds to spawn the instance on the hypervisor. [ 951.962991] env[61995]: DEBUG nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.964167] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38736866-1cc0-41d0-8b1d-3c70c14730ca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.043801] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c42e1581-193a-42bd-81e2-c68b9aa97e12 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "dc7077ac-d3fd-4e84-867a-d86328b32f5d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.463s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.149598] env[61995]: INFO nova.compute.manager [-] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Took 1.62 seconds to deallocate network for instance. [ 952.170194] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.170194] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.170194] env[61995]: DEBUG nova.network.neutron [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.170194] env[61995]: DEBUG nova.objects.instance [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lazy-loading 'info_cache' on Instance uuid 057ba9d6-4ff2-4da1-be7d-ecfadf738d74 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.285056] env[61995]: DEBUG nova.objects.instance [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.290024] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794930, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.493892] env[61995]: INFO nova.compute.manager [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Took 17.49 seconds to build instance. [ 952.615578] env[61995]: INFO nova.compute.manager [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Resuming [ 952.616387] env[61995]: DEBUG nova.objects.instance [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lazy-loading 'flavor' on Instance uuid 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.661827] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.787127] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794930, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.792827] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.792827] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.792827] env[61995]: DEBUG nova.network.neutron [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.793042] env[61995]: DEBUG nova.objects.instance [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'info_cache' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.808281] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f73ee4d-ec7c-402e-9d94-fafc4004c204 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.822281] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5de77f-c8a0-4aa0-9849-574a29e5f3e6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.855157] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdac98b2-2bfc-4883-af24-594ce24fc7f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.868375] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488d9e27-29c3-4a91-a277-4e93e8a0bc98 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.883511] env[61995]: DEBUG nova.compute.provider_tree [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.997325] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4aa56178-c687-4232-b193-ddcf81bf1a01 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.022s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.299410] env[61995]: DEBUG nova.objects.base [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 953.301101] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794930, 'name': CloneVM_Task, 'duration_secs': 1.511279} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.301470] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Created linked-clone VM from snapshot [ 953.302722] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175b01f3-d87e-42e2-8560-2d61d8f150e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.316347] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Uploading image 2b0241e5-6ffd-451c-b7b9-a9309e5284e4 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 953.349684] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 953.349684] env[61995]: value = "vm-185370" [ 953.349684] env[61995]: _type = "VirtualMachine" [ 953.349684] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 953.350239] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-61ebe9ac-cf86-455b-8b81-82a26335f762 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.366023] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease: (returnval){ [ 953.366023] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292cfc5-f53f-88e9-0455-2096a33b55f5" [ 953.366023] env[61995]: _type = "HttpNfcLease" [ 953.366023] env[61995]: } obtained for exporting VM: (result){ [ 953.366023] env[61995]: value = "vm-185370" [ 953.366023] env[61995]: _type = "VirtualMachine" [ 953.366023] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 953.366023] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the lease: (returnval){ [ 953.366023] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292cfc5-f53f-88e9-0455-2096a33b55f5" [ 953.366023] env[61995]: _type = "HttpNfcLease" [ 953.366023] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 953.376650] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.376650] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292cfc5-f53f-88e9-0455-2096a33b55f5" [ 953.376650] env[61995]: _type = "HttpNfcLease" [ 953.376650] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.389368] env[61995]: DEBUG nova.scheduler.client.report [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.632786] env[61995]: DEBUG oslo_concurrency.lockutils [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.632942] env[61995]: DEBUG oslo_concurrency.lockutils [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquired lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.636438] env[61995]: DEBUG nova.network.neutron [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 953.737232] env[61995]: DEBUG nova.network.neutron [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance_info_cache with network_info: [{"id": "0496010d-b03c-4002-b2d4-372cdbededa9", "address": "fa:16:3e:6d:e6:cd", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0496010d-b0", "ovs_interfaceid": "0496010d-b03c-4002-b2d4-372cdbededa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.878244] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.878244] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292cfc5-f53f-88e9-0455-2096a33b55f5" [ 953.878244] env[61995]: _type = "HttpNfcLease" [ 953.878244] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 953.879087] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 953.879087] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292cfc5-f53f-88e9-0455-2096a33b55f5" [ 953.879087] env[61995]: _type = "HttpNfcLease" [ 953.879087] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 953.880648] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c94660-06a1-403c-ae60-182323986827 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.889287] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961547-75a0-43fe-b235-8186030c6f59/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 953.889550] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961547-75a0-43fe-b235-8186030c6f59/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 953.960658] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.961263] env[61995]: DEBUG nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.965159] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.771s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.966900] env[61995]: INFO nova.compute.claims [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.998908] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e7b243eb-0351-4cd2-b95e-283544d4b3a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.096989] env[61995]: DEBUG nova.compute.manager [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-changed-cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 954.096989] env[61995]: DEBUG nova.compute.manager [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing instance network info cache due to event network-changed-cdaef75a-1442-4f8d-8eab-340da21a74d5. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 954.096989] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.096989] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.096989] env[61995]: DEBUG nova.network.neutron [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing network info cache for port cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 954.242104] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-057ba9d6-4ff2-4da1-be7d-ecfadf738d74" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.243035] env[61995]: DEBUG nova.objects.instance [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lazy-loading 'migration_context' on Instance uuid 057ba9d6-4ff2-4da1-be7d-ecfadf738d74 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.252420] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "847df13f-8451-49a1-9c6a-7c7c26952940" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.252836] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "847df13f-8451-49a1-9c6a-7c7c26952940" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.310975] env[61995]: DEBUG nova.network.neutron [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating instance_info_cache with network_info: [{"id": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "address": "fa:16:3e:c6:74:e1", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c280b5d-bf", "ovs_interfaceid": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.472714] env[61995]: DEBUG nova.compute.utils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 954.476327] env[61995]: DEBUG nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 954.476696] env[61995]: DEBUG nova.network.neutron [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 954.582405] env[61995]: DEBUG nova.policy [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0acd3ba123ae4a7da49803fe40ced8d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f543b47c4254f4ebeca453aea3123c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 954.585275] env[61995]: DEBUG nova.network.neutron [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [{"id": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "address": "fa:16:3e:3a:b9:92", "network": {"id": "9ecbcb9c-e75a-4a01-8abe-241d08d74461", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2006671733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0840f3a4d24ea78c35c4b273f1e413", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee71b78-02", "ovs_interfaceid": "6ee71b78-02c5-4f63-a764-d5f1bb61107c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.748140] env[61995]: DEBUG nova.objects.base [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Object Instance<057ba9d6-4ff2-4da1-be7d-ecfadf738d74> lazy-loaded attributes: info_cache,migration_context {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 954.749186] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188898c1-409e-42a7-a1e6-1b01573c2b10 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.755480] env[61995]: DEBUG nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.774586] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe57cf71-00f5-49db-b6d3-73bc9614fb42 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.782105] env[61995]: DEBUG oslo_vmware.api [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 954.782105] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52851a25-d982-aa34-3899-b8ecae2d6f7c" [ 954.782105] env[61995]: _type = "Task" [ 954.782105] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.796135] env[61995]: DEBUG oslo_vmware.api [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52851a25-d982-aa34-3899-b8ecae2d6f7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.818733] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.962394] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Acquiring lock "7d987006-afac-4973-ae69-f15b4752469b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.962394] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "7d987006-afac-4973-ae69-f15b4752469b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.962998] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Acquiring lock "7d987006-afac-4973-ae69-f15b4752469b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.962998] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "7d987006-afac-4973-ae69-f15b4752469b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.962998] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "7d987006-afac-4973-ae69-f15b4752469b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.966081] env[61995]: INFO nova.compute.manager [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Terminating instance [ 954.968507] env[61995]: DEBUG nova.compute.manager [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.968906] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 954.970926] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3b947e2-ffe7-421a-af98-c6b8160bb74e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.978151] env[61995]: DEBUG nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.981820] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 954.981820] env[61995]: value = "task-794932" [ 954.981820] env[61995]: _type = "Task" [ 954.981820] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.995704] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794932, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.061916] env[61995]: DEBUG nova.network.neutron [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updated VIF entry in instance network info cache for port cdaef75a-1442-4f8d-8eab-340da21a74d5. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 955.062357] env[61995]: DEBUG nova.network.neutron [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.093381] env[61995]: DEBUG oslo_concurrency.lockutils [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Releasing lock "refresh_cache-0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.093381] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f81449a-fc9b-4b6b-8337-28655bea0aa8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.103110] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Resuming the VM {{(pid=61995) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 955.103806] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96c90bb9-9600-47cf-b68b-cc86bae29b42 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.118141] env[61995]: DEBUG oslo_vmware.api [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 955.118141] env[61995]: value = "task-794933" [ 955.118141] env[61995]: _type = "Task" [ 955.118141] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.127135] env[61995]: DEBUG oslo_vmware.api [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.192372] env[61995]: DEBUG nova.network.neutron [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Successfully created port: 2327f433-5c63-4d83-8050-6661d9ad60eb {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.294649] env[61995]: DEBUG oslo_vmware.api [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52851a25-d982-aa34-3899-b8ecae2d6f7c, 'name': SearchDatastore_Task, 'duration_secs': 0.014063} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.296079] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.299021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.309652] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3136b1ac-e8df-4ec5-9d62-02c997e43568 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.320067] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2d5db4-3d61-446c-b7f3-b9d578ecb1a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.324688] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 955.325605] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7146bc2b-7160-4fd0-a1fb-11118d1089c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.363616] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275d69ba-14ec-4816-991e-eadbeb31c5cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.366520] env[61995]: DEBUG oslo_vmware.api [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 955.366520] env[61995]: value = "task-794934" [ 955.366520] env[61995]: _type = "Task" [ 955.366520] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.374828] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c45abdc-7bfa-4596-a9ac-de9f43b07e3c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.383035] env[61995]: DEBUG oslo_vmware.api [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794934, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.397060] env[61995]: DEBUG nova.compute.provider_tree [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.496916] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794932, 'name': PowerOffVM_Task, 'duration_secs': 0.279515} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.497294] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 955.497596] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 955.499948] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185340', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'name': 'volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d987006-afac-4973-ae69-f15b4752469b', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'serial': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 955.499948] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9eaf49-a1d3-43f6-aac8-255008367931 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.525161] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a967fe1a-dff0-471f-9284-1ade4ec9e68c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.536219] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e4724f-90f7-4883-87f6-f088875f093c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.559210] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a469bd5-8fbb-4468-9c51-aea7fb3a28f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.576382] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d5ec89b-3791-4b4c-93c5-22e31544951d req-a6bc8d31-b89d-496c-9079-4d24f597235c service nova] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.576791] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] The volume has not been displaced from its original location: [datastore1] volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02/volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 955.582571] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Reconfiguring VM instance instance-0000004a to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 955.582978] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab184ff6-1ae3-49af-8f5b-0d5b8fea6bb5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.605376] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 955.605376] env[61995]: value = "task-794935" [ 955.605376] env[61995]: _type = "Task" [ 955.605376] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.615676] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.627864] env[61995]: DEBUG oslo_vmware.api [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794933, 'name': PowerOnVM_Task} progress is 93%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.669621] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.670050] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.878967] env[61995]: DEBUG oslo_vmware.api [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-794934, 'name': PowerOnVM_Task, 'duration_secs': 0.54296} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.879542] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 955.880117] env[61995]: DEBUG nova.compute.manager [None req-2f8bb786-a195-480b-aed1-11e1c44f5432 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.881027] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7640f367-855b-4ebd-a864-007857c3595f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.900942] env[61995]: DEBUG nova.scheduler.client.report [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.991956] env[61995]: DEBUG nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.019480] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.019832] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.020092] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.020351] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.020531] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.020803] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.021039] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.021218] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.021394] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.021559] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.021745] env[61995]: DEBUG nova.virt.hardware [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.022669] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f828558-c51a-48ac-9f6c-4d6c054747a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.033894] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20e0764-b37c-4c7c-b030-1245161c5109 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.116805] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794935, 'name': ReconfigVM_Task, 'duration_secs': 0.246334} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.117494] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Reconfigured VM instance instance-0000004a to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 956.122721] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8998c53-c4c1-490e-aca4-c76efdaa512a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.146080] env[61995]: DEBUG oslo_vmware.api [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794933, 'name': PowerOnVM_Task, 'duration_secs': 0.733399} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.147566] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Resumed the VM {{(pid=61995) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 956.147888] env[61995]: DEBUG nova.compute.manager [None req-660b02fd-a4b5-4785-8ddf-00f0161b08fb tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.148363] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 956.148363] env[61995]: value = "task-794936" [ 956.148363] env[61995]: _type = "Task" [ 956.148363] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.149398] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f660f854-1cf6-45b2-af8f-8776538eb237 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.167143] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794936, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.173027] env[61995]: DEBUG nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.406505] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.407745] env[61995]: DEBUG nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 956.412511] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.775s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.413279] env[61995]: DEBUG nova.objects.instance [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'resources' on Instance uuid 88638f35-950d-42b4-92b2-ed0dced9148c {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.663627] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794936, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.697245] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.792695] env[61995]: DEBUG nova.compute.manager [req-55f3297b-8651-4ad3-9105-fd6f7347ede2 req-45237ddc-843c-4e82-ab04-78d35dc50ae3 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Received event network-vif-plugged-2327f433-5c63-4d83-8050-6661d9ad60eb {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 956.792932] env[61995]: DEBUG oslo_concurrency.lockutils [req-55f3297b-8651-4ad3-9105-fd6f7347ede2 req-45237ddc-843c-4e82-ab04-78d35dc50ae3 service nova] Acquiring lock "24b974c0-9567-46a3-96f6-fac1e602f46e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.793234] env[61995]: DEBUG oslo_concurrency.lockutils [req-55f3297b-8651-4ad3-9105-fd6f7347ede2 req-45237ddc-843c-4e82-ab04-78d35dc50ae3 service nova] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.793448] env[61995]: DEBUG oslo_concurrency.lockutils [req-55f3297b-8651-4ad3-9105-fd6f7347ede2 req-45237ddc-843c-4e82-ab04-78d35dc50ae3 service nova] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.793609] env[61995]: DEBUG nova.compute.manager [req-55f3297b-8651-4ad3-9105-fd6f7347ede2 req-45237ddc-843c-4e82-ab04-78d35dc50ae3 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] No waiting events found dispatching network-vif-plugged-2327f433-5c63-4d83-8050-6661d9ad60eb {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.793781] env[61995]: WARNING nova.compute.manager [req-55f3297b-8651-4ad3-9105-fd6f7347ede2 req-45237ddc-843c-4e82-ab04-78d35dc50ae3 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Received unexpected event network-vif-plugged-2327f433-5c63-4d83-8050-6661d9ad60eb for instance with vm_state building and task_state spawning. [ 956.906160] env[61995]: DEBUG nova.network.neutron [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Successfully updated port: 2327f433-5c63-4d83-8050-6661d9ad60eb {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.914555] env[61995]: DEBUG nova.compute.utils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.916548] env[61995]: DEBUG nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.916822] env[61995]: DEBUG nova.network.neutron [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 956.988167] env[61995]: DEBUG nova.policy [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0acd3ba123ae4a7da49803fe40ced8d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f543b47c4254f4ebeca453aea3123c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 957.158688] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757269f5-1772-4bdb-b982-718e05e6dac2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.167802] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794936, 'name': ReconfigVM_Task, 'duration_secs': 0.789316} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.169458] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185340', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'name': 'volume-e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d987006-afac-4973-ae69-f15b4752469b', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02', 'serial': 'e9de6b68-7402-4a32-b694-1d36fc3a9a02'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 957.169753] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 957.170580] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24cbee9-9120-41e3-8895-5dc34cd2017c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.173861] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4ca456-22a3-48f3-b4f9-e810d56913ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.184428] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 957.208274] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc8a898d-726c-4d52-a6a6-e0f37513b075 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.213021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906b9d46-605d-4f85-ac05-0b7309a88a7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.219498] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62861f0b-7618-4571-8989-4b96df989c89 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.234642] env[61995]: DEBUG nova.compute.provider_tree [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.336515] env[61995]: DEBUG nova.network.neutron [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Successfully created port: d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.409308] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.409636] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.409636] env[61995]: DEBUG nova.network.neutron [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.423974] env[61995]: DEBUG nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 957.737409] env[61995]: DEBUG nova.scheduler.client.report [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.960856] env[61995]: DEBUG nova.network.neutron [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 958.176504] env[61995]: DEBUG nova.network.neutron [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updating instance_info_cache with network_info: [{"id": "2327f433-5c63-4d83-8050-6661d9ad60eb", "address": "fa:16:3e:d0:20:ad", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2327f433-5c", "ovs_interfaceid": "2327f433-5c63-4d83-8050-6661d9ad60eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.242369] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.245956] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.584s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.246076] env[61995]: DEBUG nova.objects.instance [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lazy-loading 'resources' on Instance uuid 5a6bd8a7-2633-42d2-b853-4e3a5690c77e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.276674] env[61995]: INFO nova.scheduler.client.report [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted allocations for instance 88638f35-950d-42b4-92b2-ed0dced9148c [ 958.432785] env[61995]: DEBUG nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 958.461178] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.461448] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.461610] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.461798] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.461948] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.462116] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.462360] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.462728] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.462728] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.462941] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.463070] env[61995]: DEBUG nova.virt.hardware [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.463922] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8ad724-3d2f-4de3-a06f-e422902fabfc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.474538] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661d1039-6f3d-4614-8d77-14fcbe0ef7d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.680692] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.681089] env[61995]: DEBUG nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Instance network_info: |[{"id": "2327f433-5c63-4d83-8050-6661d9ad60eb", "address": "fa:16:3e:d0:20:ad", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2327f433-5c", "ovs_interfaceid": "2327f433-5c63-4d83-8050-6661d9ad60eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.681581] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:20:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2327f433-5c63-4d83-8050-6661d9ad60eb', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.692924] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating folder: Project (3f543b47c4254f4ebeca453aea3123c8). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 958.692924] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b175d490-e555-4e21-8866-ccaddb9d4c00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.707129] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created folder: Project (3f543b47c4254f4ebeca453aea3123c8) in parent group-v185203. [ 958.707385] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating folder: Instances. Parent ref: group-v185371. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 958.707656] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f56271c5-f048-4689-8b24-f4cad9defccf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.720226] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created folder: Instances in parent group-v185371. [ 958.720561] env[61995]: DEBUG oslo.service.loopingcall [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.720788] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 958.721014] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9539058-39ff-4358-bbc6-dcb4518acd42 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.743148] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.743148] env[61995]: value = "task-794940" [ 958.743148] env[61995]: _type = "Task" [ 958.743148] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.754353] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794940, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.789359] env[61995]: DEBUG oslo_concurrency.lockutils [None req-85034a7c-86a8-4b41-8a5b-80256ab29a56 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "88638f35-950d-42b4-92b2-ed0dced9148c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.200s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.938364] env[61995]: DEBUG nova.compute.manager [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Received event network-changed-2327f433-5c63-4d83-8050-6661d9ad60eb {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.938364] env[61995]: DEBUG nova.compute.manager [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Refreshing instance network info cache due to event network-changed-2327f433-5c63-4d83-8050-6661d9ad60eb. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 958.938364] env[61995]: DEBUG oslo_concurrency.lockutils [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] Acquiring lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.939736] env[61995]: DEBUG oslo_concurrency.lockutils [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] Acquired lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.940199] env[61995]: DEBUG nova.network.neutron [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Refreshing network info cache for port 2327f433-5c63-4d83-8050-6661d9ad60eb {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.003176] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e641a9-8d8a-49b3-b294-68c2ffbd87c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.013467] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daea9bec-e218-42ad-a5ba-2d3e80de61f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.051740] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ea781d-a310-41b5-a277-e9b969eb5d6e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.063015] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc712c9b-5ace-4304-8de0-ae3ed9b47040 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.082071] env[61995]: DEBUG nova.compute.provider_tree [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.126773] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 959.126981] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 959.128199] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Deleting the datastore file [datastore1] 7d987006-afac-4973-ae69-f15b4752469b {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.128199] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db7ee5fc-454e-4a7e-ba18-60dacc24d1b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.145602] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for the task: (returnval){ [ 959.145602] env[61995]: value = "task-794941" [ 959.145602] env[61995]: _type = "Task" [ 959.145602] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.154494] env[61995]: DEBUG oslo_concurrency.lockutils [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.154780] env[61995]: DEBUG oslo_concurrency.lockutils [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.166086] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794941, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.177904] env[61995]: DEBUG nova.network.neutron [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Successfully updated port: d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.255436] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794940, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.583803] env[61995]: DEBUG nova.scheduler.client.report [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.658227] env[61995]: DEBUG oslo_vmware.api [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Task: {'id': task-794941, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135044} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.658443] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.658504] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 959.659049] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 959.659049] env[61995]: INFO nova.compute.manager [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Took 4.69 seconds to destroy the instance on the hypervisor. [ 959.659212] env[61995]: DEBUG oslo.service.loopingcall [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.659316] env[61995]: DEBUG nova.compute.manager [-] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.659419] env[61995]: DEBUG nova.network.neutron [-] [instance: 7d987006-afac-4973-ae69-f15b4752469b] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 959.662421] env[61995]: DEBUG nova.compute.utils [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 959.682538] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.682726] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.683110] env[61995]: DEBUG nova.network.neutron [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 959.755746] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794940, 'name': CreateVM_Task, 'duration_secs': 0.520785} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.756521] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 959.756807] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.756908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.757295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.757607] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a877c40a-106b-42f8-9dab-777f6027143a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.764087] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 959.764087] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5277190c-c34b-82da-b783-bdf94704164f" [ 959.764087] env[61995]: _type = "Task" [ 959.764087] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.774197] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5277190c-c34b-82da-b783-bdf94704164f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.853444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "8ae65d63-16c7-4c67-a8c2-73968f63457d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.853685] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.089223] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.091832] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.796s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.093428] env[61995]: INFO nova.compute.claims [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.114757] env[61995]: DEBUG nova.network.neutron [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updated VIF entry in instance network info cache for port 2327f433-5c63-4d83-8050-6661d9ad60eb. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.115315] env[61995]: DEBUG nova.network.neutron [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updating instance_info_cache with network_info: [{"id": "2327f433-5c63-4d83-8050-6661d9ad60eb", "address": "fa:16:3e:d0:20:ad", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2327f433-5c", "ovs_interfaceid": "2327f433-5c63-4d83-8050-6661d9ad60eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.118447] env[61995]: INFO nova.scheduler.client.report [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Deleted allocations for instance 5a6bd8a7-2633-42d2-b853-4e3a5690c77e [ 960.166011] env[61995]: DEBUG oslo_concurrency.lockutils [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.239249] env[61995]: DEBUG nova.network.neutron [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 960.278555] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5277190c-c34b-82da-b783-bdf94704164f, 'name': SearchDatastore_Task, 'duration_secs': 0.015403} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.281636] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.281911] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.282258] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.282351] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.282568] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.282870] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97de8001-e55b-4d3f-8a98-6479eb90582c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.294231] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.294453] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 960.295315] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48676f86-fb80-486f-8138-7e762c7cff37 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.302310] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 960.302310] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5211cbb1-0d56-e2e6-59d7-643060fd18ef" [ 960.302310] env[61995]: _type = "Task" [ 960.302310] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.312870] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5211cbb1-0d56-e2e6-59d7-643060fd18ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.358124] env[61995]: DEBUG nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 960.536518] env[61995]: DEBUG nova.network.neutron [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Updating instance_info_cache with network_info: [{"id": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "address": "fa:16:3e:58:1b:ff", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8ff3c80-bf", "ovs_interfaceid": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.622956] env[61995]: DEBUG oslo_concurrency.lockutils [req-d192ae70-dd1a-4066-89c3-32c4c9e69bc3 req-eaac868d-9759-4488-b07d-f761c230bf38 service nova] Releasing lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.626498] env[61995]: DEBUG oslo_concurrency.lockutils [None req-655499ab-03a5-43a1-8ffd-ef3cfab139af tempest-ServersTestFqdnHostnames-1717723339 tempest-ServersTestFqdnHostnames-1717723339-project-member] Lock "5a6bd8a7-2633-42d2-b853-4e3a5690c77e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.755s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.817761] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5211cbb1-0d56-e2e6-59d7-643060fd18ef, 'name': SearchDatastore_Task, 'duration_secs': 0.017589} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.818801] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a9c8c09-c76b-406a-ab6f-005a703fadf5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.825953] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 960.825953] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526a7b0d-c3de-f3cb-a323-61f09ea2b268" [ 960.825953] env[61995]: _type = "Task" [ 960.825953] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.835189] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526a7b0d-c3de-f3cb-a323-61f09ea2b268, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.880444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.980705] env[61995]: DEBUG nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Received event network-vif-plugged-d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.980931] env[61995]: DEBUG oslo_concurrency.lockutils [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] Acquiring lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.981256] env[61995]: DEBUG oslo_concurrency.lockutils [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.981482] env[61995]: DEBUG oslo_concurrency.lockutils [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.981688] env[61995]: DEBUG nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] No waiting events found dispatching network-vif-plugged-d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.981890] env[61995]: WARNING nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Received unexpected event network-vif-plugged-d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 for instance with vm_state building and task_state spawning. [ 960.982114] env[61995]: DEBUG nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Received event network-changed-d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.982432] env[61995]: DEBUG nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Refreshing instance network info cache due to event network-changed-d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 960.982609] env[61995]: DEBUG oslo_concurrency.lockutils [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] Acquiring lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.039345] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.040501] env[61995]: DEBUG nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Instance network_info: |[{"id": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "address": "fa:16:3e:58:1b:ff", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8ff3c80-bf", "ovs_interfaceid": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.040966] env[61995]: DEBUG oslo_concurrency.lockutils [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] Acquired lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.045049] env[61995]: DEBUG nova.network.neutron [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Refreshing network info cache for port d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.046582] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:1b:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.059985] env[61995]: DEBUG oslo.service.loopingcall [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.061390] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 961.061629] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b837ced8-3bd1-477a-b534-ac0dafb756ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.085840] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.085840] env[61995]: value = "task-794942" [ 961.085840] env[61995]: _type = "Task" [ 961.085840] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.095619] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794942, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.223010] env[61995]: DEBUG oslo_concurrency.lockutils [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.223322] env[61995]: DEBUG oslo_concurrency.lockutils [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.223560] env[61995]: INFO nova.compute.manager [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Attaching volume caeb58b4-c8a5-46b3-9f21-57540bd5b068 to /dev/sdb [ 961.269786] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4616ca1-46a6-4457-8b3a-61597bc760e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.295612] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b648b23-ef25-4ba6-b555-ea06d193d452 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.299066] env[61995]: DEBUG nova.network.neutron [-] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.326031] env[61995]: DEBUG nova.virt.block_device [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updating existing volume attachment record: 0d95a577-69cc-4c3a-9964-269013087241 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 961.339778] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526a7b0d-c3de-f3cb-a323-61f09ea2b268, 'name': SearchDatastore_Task, 'duration_secs': 0.010064} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.339929] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.340212] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/24b974c0-9567-46a3-96f6-fac1e602f46e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 961.340525] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebd2a291-be2b-4715-bac6-0272cdb22318 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.349976] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 961.349976] env[61995]: value = "task-794943" [ 961.349976] env[61995]: _type = "Task" [ 961.349976] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.360420] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.422741] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8055532-6a9d-49a5-b570-270ce14b9edc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.433025] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc9031a-524f-4038-8c88-b42b03bc1375 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.463296] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a021ac8-fe50-408c-ada8-8aca1b286cec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.473548] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1d9879-6337-40bf-84a9-46d2cf5bff68 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.489420] env[61995]: DEBUG nova.compute.provider_tree [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.599162] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794942, 'name': CreateVM_Task, 'duration_secs': 0.497247} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.602565] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 961.606023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.606023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.606023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 961.606023] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc1dda67-91d3-45cc-90bc-cb6465b2578e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.610206] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 961.610206] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5224f168-ff9c-f779-eb66-b1799e0b3146" [ 961.610206] env[61995]: _type = "Task" [ 961.610206] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.619698] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5224f168-ff9c-f779-eb66-b1799e0b3146, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.806378] env[61995]: INFO nova.compute.manager [-] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Took 2.15 seconds to deallocate network for instance. [ 961.842993] env[61995]: DEBUG nova.network.neutron [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Updated VIF entry in instance network info cache for port d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 961.843693] env[61995]: DEBUG nova.network.neutron [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Updating instance_info_cache with network_info: [{"id": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "address": "fa:16:3e:58:1b:ff", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8ff3c80-bf", "ovs_interfaceid": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.861280] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794943, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.992631] env[61995]: DEBUG nova.scheduler.client.report [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.123563] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5224f168-ff9c-f779-eb66-b1799e0b3146, 'name': SearchDatastore_Task, 'duration_secs': 0.077527} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.124010] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.124360] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.124708] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.124918] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.125208] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.125570] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f005132-2d5b-4abf-81cc-e5ec8a726eb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.135334] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.135633] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 962.136663] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93785a7b-0a21-475d-a599-00ef9a0c7f45 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.143304] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 962.143304] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52693b79-6e62-e96b-b602-21df97f57a16" [ 962.143304] env[61995]: _type = "Task" [ 962.143304] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.154150] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52693b79-6e62-e96b-b602-21df97f57a16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.347905] env[61995]: DEBUG oslo_concurrency.lockutils [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] Releasing lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.348292] env[61995]: DEBUG nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Received event network-vif-deleted-a29a835c-8bc8-4c90-b691-44a040b5c90f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.348490] env[61995]: INFO nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Neutron deleted interface a29a835c-8bc8-4c90-b691-44a040b5c90f; detaching it from the instance and deleting it from the info cache [ 962.348671] env[61995]: DEBUG nova.network.neutron [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.360853] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581472} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.361231] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/24b974c0-9567-46a3-96f6-fac1e602f46e.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 962.361376] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 962.361593] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f6b548f-6a52-44f7-a501-9dc22d4fd710 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.369701] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 962.369701] env[61995]: value = "task-794947" [ 962.369701] env[61995]: _type = "Task" [ 962.369701] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.374371] env[61995]: INFO nova.compute.manager [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Took 0.57 seconds to detach 1 volumes for instance. [ 962.377733] env[61995]: DEBUG nova.compute.manager [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Deleting volume: e9de6b68-7402-4a32-b694-1d36fc3a9a02 {{(pid=61995) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 962.383213] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794947, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.498171] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.498715] env[61995]: DEBUG nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 962.502044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.203s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.656153] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52693b79-6e62-e96b-b602-21df97f57a16, 'name': SearchDatastore_Task, 'duration_secs': 0.00907} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.657768] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1d1269f-e676-4a24-820b-72e053aecec5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.662998] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 962.662998] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52521445-adaa-a46c-83cb-d15738835124" [ 962.662998] env[61995]: _type = "Task" [ 962.662998] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.671299] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52521445-adaa-a46c-83cb-d15738835124, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.852020] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdb15c03-4215-47eb-82a2-86b82a023870 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.861536] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba90e449-dfd6-4ee7-a22b-192a1d83927c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.888243] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794947, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081921} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.897893] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.898398] env[61995]: DEBUG nova.compute.manager [req-0372b02f-5d22-4c98-bdae-31883b247485 req-e6417334-4dac-4dcb-a752-ffad4ae7f2a6 service nova] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Detach interface failed, port_id=a29a835c-8bc8-4c90-b691-44a040b5c90f, reason: Instance 7d987006-afac-4973-ae69-f15b4752469b could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 962.899381] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a91c49-413d-498f-86f0-f702c2ea477e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.923152] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/24b974c0-9567-46a3-96f6-fac1e602f46e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.924389] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b6bdd3e-2a4b-4456-949f-5e49a03dbe65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.941729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.944265] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 962.944265] env[61995]: value = "task-794949" [ 962.944265] env[61995]: _type = "Task" [ 962.944265] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.955787] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.007619] env[61995]: DEBUG nova.compute.utils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.012084] env[61995]: DEBUG nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.012720] env[61995]: DEBUG nova.network.neutron [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 963.063587] env[61995]: DEBUG nova.policy [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '92df3999e9594fba9444e485dceb93a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978c59760428439cb83c95c9a39be6e6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 963.177808] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52521445-adaa-a46c-83cb-d15738835124, 'name': SearchDatastore_Task, 'duration_secs': 0.010999} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.177808] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.178058] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/5cb4c1d9-e661-48b4-8600-1b823d63ddbb.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 963.178421] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c372f777-df1a-4201-9c2a-1b92a2a824b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.186978] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 963.186978] env[61995]: value = "task-794950" [ 963.186978] env[61995]: _type = "Task" [ 963.186978] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.198702] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794950, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.295336] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adbf0879-fa2b-46b2-b26e-7a1e8e0ad187 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.305238] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eecc2ddd-d25a-4aa4-b226-4ca597f373e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.341272] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e60270-cd12-4061-b13b-c40a65e0c55e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.349915] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8779082b-2112-4123-9106-42d238fbdc34 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.364195] env[61995]: DEBUG nova.compute.provider_tree [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.454549] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794949, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.517879] env[61995]: DEBUG nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 963.574481] env[61995]: DEBUG nova.network.neutron [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Successfully created port: 4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.700310] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794950, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.867719] env[61995]: DEBUG nova.scheduler.client.report [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.955392] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794949, 'name': ReconfigVM_Task, 'duration_secs': 0.556796} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.955720] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/24b974c0-9567-46a3-96f6-fac1e602f46e.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.956420] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95e504f7-bcd4-4bd2-a198-46f2d2100de1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.962754] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 963.962754] env[61995]: value = "task-794952" [ 963.962754] env[61995]: _type = "Task" [ 963.962754] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.971407] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794952, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.183437] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.183685] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.183898] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.184109] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.184298] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.188622] env[61995]: INFO nova.compute.manager [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Terminating instance [ 964.195299] env[61995]: DEBUG nova.compute.manager [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.195299] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 964.196072] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e23d301-06a6-4b67-a89b-5fc8f4c1eef1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.205204] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 964.208271] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fe0566b-1ccc-4a3a-9b12-e60a6678a99c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.211146] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794950, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.665487} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.211436] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/5cb4c1d9-e661-48b4-8600-1b823d63ddbb.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 964.211640] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.212718] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca7452b8-c599-495c-bd44-04bcbf4827e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.218586] env[61995]: DEBUG oslo_vmware.api [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 964.218586] env[61995]: value = "task-794953" [ 964.218586] env[61995]: _type = "Task" [ 964.218586] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.224129] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 964.224129] env[61995]: value = "task-794954" [ 964.224129] env[61995]: _type = "Task" [ 964.224129] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.233735] env[61995]: DEBUG oslo_vmware.api [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.239234] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794954, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.473944] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794952, 'name': Rename_Task, 'duration_secs': 0.171159} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.475800] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 964.475800] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54ddf48d-0635-4fd8-b2ea-32d1ed7a36f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.486624] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 964.486624] env[61995]: value = "task-794955" [ 964.486624] env[61995]: _type = "Task" [ 964.486624] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.502763] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.530032] env[61995]: DEBUG nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 964.565035] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.565035] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.565035] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.565370] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.565529] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.565715] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.566110] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.566300] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.566509] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.567251] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.567351] env[61995]: DEBUG nova.virt.hardware [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.568320] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d2e491-a59e-4fab-b0d9-9bec265362b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.576948] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426e86fc-69b7-4696-b846-aa923547d94a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.730241] env[61995]: DEBUG oslo_vmware.api [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794953, 'name': PowerOffVM_Task, 'duration_secs': 0.209066} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.732050] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 964.732283] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 964.732557] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bb3b344-e8e1-4276-9498-4ac1971e54d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.738984] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794954, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086464} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.739287] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 964.740099] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5570580-1da2-409a-a071-905272fcc26e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.762462] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/5cb4c1d9-e661-48b4-8600-1b823d63ddbb.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.762829] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d54f046-e480-4e67-a115-0454b432c791 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.781740] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 964.781740] env[61995]: value = "task-794957" [ 964.781740] env[61995]: _type = "Task" [ 964.781740] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.793089] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.802328] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 964.802684] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 964.802783] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleting the datastore file [datastore2] 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.803069] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0b71f36-65df-4fce-b639-b3cda00609c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.809339] env[61995]: DEBUG oslo_vmware.api [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for the task: (returnval){ [ 964.809339] env[61995]: value = "task-794958" [ 964.809339] env[61995]: _type = "Task" [ 964.809339] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.817700] env[61995]: DEBUG oslo_vmware.api [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.883934] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.382s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.886803] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.190s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.888395] env[61995]: INFO nova.compute.claims [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.898961] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961547-75a0-43fe-b235-8186030c6f59/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 964.899888] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309e8e4f-85d4-46c4-bf33-d57f124568da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.906221] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961547-75a0-43fe-b235-8186030c6f59/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 964.906399] env[61995]: ERROR oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961547-75a0-43fe-b235-8186030c6f59/disk-0.vmdk due to incomplete transfer. [ 964.906629] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9ce80075-9c5e-4143-80f5-6c6105aabadb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.914566] env[61995]: DEBUG oslo_vmware.rw_handles [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961547-75a0-43fe-b235-8186030c6f59/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 964.914778] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Uploaded image 2b0241e5-6ffd-451c-b7b9-a9309e5284e4 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 964.917299] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 964.917629] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-07fbd05b-5fb4-40c0-add2-f6971aaded3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.925914] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 964.925914] env[61995]: value = "task-794959" [ 964.925914] env[61995]: _type = "Task" [ 964.925914] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.935012] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794959, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.996754] env[61995]: DEBUG oslo_vmware.api [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794955, 'name': PowerOnVM_Task, 'duration_secs': 0.508407} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.997419] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 964.997652] env[61995]: INFO nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Took 9.01 seconds to spawn the instance on the hypervisor. [ 964.997856] env[61995]: DEBUG nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.999033] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65372c2-634f-4e1a-88da-b5b799d48ca5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.233344] env[61995]: DEBUG nova.compute.manager [req-fff06ab1-3d0e-4c0a-ada4-83927483bcf9 req-18423361-1114-4c5a-b632-fa15cbf9d261 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Received event network-vif-plugged-4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 965.233582] env[61995]: DEBUG oslo_concurrency.lockutils [req-fff06ab1-3d0e-4c0a-ada4-83927483bcf9 req-18423361-1114-4c5a-b632-fa15cbf9d261 service nova] Acquiring lock "847df13f-8451-49a1-9c6a-7c7c26952940-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.233922] env[61995]: DEBUG oslo_concurrency.lockutils [req-fff06ab1-3d0e-4c0a-ada4-83927483bcf9 req-18423361-1114-4c5a-b632-fa15cbf9d261 service nova] Lock "847df13f-8451-49a1-9c6a-7c7c26952940-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.234025] env[61995]: DEBUG oslo_concurrency.lockutils [req-fff06ab1-3d0e-4c0a-ada4-83927483bcf9 req-18423361-1114-4c5a-b632-fa15cbf9d261 service nova] Lock "847df13f-8451-49a1-9c6a-7c7c26952940-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.234369] env[61995]: DEBUG nova.compute.manager [req-fff06ab1-3d0e-4c0a-ada4-83927483bcf9 req-18423361-1114-4c5a-b632-fa15cbf9d261 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] No waiting events found dispatching network-vif-plugged-4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.234579] env[61995]: WARNING nova.compute.manager [req-fff06ab1-3d0e-4c0a-ada4-83927483bcf9 req-18423361-1114-4c5a-b632-fa15cbf9d261 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Received unexpected event network-vif-plugged-4554d960-3f34-464e-abd8-8f16166194b3 for instance with vm_state building and task_state spawning. [ 965.297723] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.319884] env[61995]: DEBUG oslo_vmware.api [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Task: {'id': task-794958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.348956} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.320203] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.320398] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 965.320588] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 965.320765] env[61995]: INFO nova.compute.manager [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 965.321024] env[61995]: DEBUG oslo.service.loopingcall [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.321229] env[61995]: DEBUG nova.compute.manager [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.321328] env[61995]: DEBUG nova.network.neutron [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 965.437724] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794959, 'name': Destroy_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.479535] env[61995]: INFO nova.scheduler.client.report [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted allocation for migration f6f8d556-5345-4a35-8742-c2e7db554c68 [ 965.517166] env[61995]: INFO nova.compute.manager [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Took 17.04 seconds to build instance. [ 965.767989] env[61995]: DEBUG nova.compute.manager [req-5728d7be-bcf1-403c-a623-39b9e24a189b req-3065af82-e8fa-4b40-922a-a78e9b2a4b55 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Received event network-vif-deleted-6ee71b78-02c5-4f63-a764-d5f1bb61107c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 965.768283] env[61995]: INFO nova.compute.manager [req-5728d7be-bcf1-403c-a623-39b9e24a189b req-3065af82-e8fa-4b40-922a-a78e9b2a4b55 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Neutron deleted interface 6ee71b78-02c5-4f63-a764-d5f1bb61107c; detaching it from the instance and deleting it from the info cache [ 965.768501] env[61995]: DEBUG nova.network.neutron [req-5728d7be-bcf1-403c-a623-39b9e24a189b req-3065af82-e8fa-4b40-922a-a78e9b2a4b55 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.787741] env[61995]: DEBUG nova.network.neutron [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Successfully updated port: 4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.797370] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794957, 'name': ReconfigVM_Task, 'duration_secs': 0.584146} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.797726] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/5cb4c1d9-e661-48b4-8600-1b823d63ddbb.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.798512] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50e485e1-ef0a-4bee-bcca-bb5b9992b625 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.807186] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 965.807186] env[61995]: value = "task-794960" [ 965.807186] env[61995]: _type = "Task" [ 965.807186] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.816044] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794960, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.879972] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 965.880247] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 965.881115] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a617cbfc-e36e-441a-a0ab-d66d965d4b85 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.898662] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4591c0d-cfde-4b88-af6c-24c300ac7aca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.926951] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068/volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.927580] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6500d546-dab3-4ef6-8309-118371e5ebf7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.952337] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794959, 'name': Destroy_Task, 'duration_secs': 0.62411} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.953834] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Destroyed the VM [ 965.954099] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 965.954509] env[61995]: DEBUG oslo_vmware.api [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 965.954509] env[61995]: value = "task-794961" [ 965.954509] env[61995]: _type = "Task" [ 965.954509] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.954732] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4bbd5b94-ca09-47b9-8e4d-4846c5625b91 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.967617] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 965.967617] env[61995]: value = "task-794962" [ 965.967617] env[61995]: _type = "Task" [ 965.967617] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.970969] env[61995]: DEBUG oslo_vmware.api [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794961, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.983578] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794962, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.990913] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2220f506-84d1-4427-907a-45d6551bb3c8 tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.397s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.024806] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb4fca76-a6b6-4933-8112-bab5bcc09b53 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.556s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.078099] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.078099] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.078264] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.078431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.078650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.080949] env[61995]: INFO nova.compute.manager [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Terminating instance [ 966.084271] env[61995]: DEBUG nova.compute.manager [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.084271] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 966.085262] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c110f98-80e1-4df0-81f1-6ce5148229ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.096956] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.097427] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb2bf261-a1d4-47a6-beec-952fdb19bbdc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.106775] env[61995]: DEBUG oslo_vmware.api [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 966.106775] env[61995]: value = "task-794963" [ 966.106775] env[61995]: _type = "Task" [ 966.106775] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.121621] env[61995]: DEBUG oslo_vmware.api [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794963, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.135842] env[61995]: DEBUG nova.network.neutron [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.143676] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350e28c7-d49e-4bf5-93b9-b2b07c709b71 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.153606] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71bf4e7-ad47-44d4-9f53-78117cd3e038 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.190099] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedad99b-e434-4a38-abab-4a142f95fc4e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.199260] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a87a07-b4ea-4c1e-aefa-ccb62bf92f02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.214309] env[61995]: DEBUG nova.compute.provider_tree [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.271643] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5668f49-471f-4d84-ab08-5b941556888a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.282856] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1343f9e9-9b61-43b1-a6ce-3f17a102bc12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.295439] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.295643] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquired lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.295799] env[61995]: DEBUG nova.network.neutron [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 966.321742] env[61995]: DEBUG nova.compute.manager [req-5728d7be-bcf1-403c-a623-39b9e24a189b req-3065af82-e8fa-4b40-922a-a78e9b2a4b55 service nova] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Detach interface failed, port_id=6ee71b78-02c5-4f63-a764-d5f1bb61107c, reason: Instance 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 966.329089] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794960, 'name': Rename_Task, 'duration_secs': 0.281293} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.329463] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 966.329735] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a81e8bb7-ea8e-4875-830a-583e214964cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.336529] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 966.336529] env[61995]: value = "task-794964" [ 966.336529] env[61995]: _type = "Task" [ 966.336529] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.354434] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794964, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.466341] env[61995]: DEBUG oslo_vmware.api [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794961, 'name': ReconfigVM_Task, 'duration_secs': 0.41108} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.466646] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfigured VM instance instance-0000004c to attach disk [datastore2] volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068/volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.471855] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40c1cfb2-74d1-450d-a294-c6036fe65f37 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.491814] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794962, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.493435] env[61995]: DEBUG oslo_vmware.api [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 966.493435] env[61995]: value = "task-794965" [ 966.493435] env[61995]: _type = "Task" [ 966.493435] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.501430] env[61995]: DEBUG oslo_vmware.api [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794965, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.616915] env[61995]: DEBUG oslo_vmware.api [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794963, 'name': PowerOffVM_Task, 'duration_secs': 0.264605} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.617660] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.617660] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 966.617901] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07eb64db-e142-4e26-ad86-9619edc51939 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.639400] env[61995]: INFO nova.compute.manager [-] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Took 1.32 seconds to deallocate network for instance. [ 966.693776] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.694037] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.705450] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 966.705705] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 966.705903] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleting the datastore file [datastore1] 057ba9d6-4ff2-4da1-be7d-ecfadf738d74 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.706995] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b2519c1-34ae-4783-a600-640c91cbf4e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.717256] env[61995]: DEBUG nova.scheduler.client.report [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.721337] env[61995]: DEBUG oslo_vmware.api [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 966.721337] env[61995]: value = "task-794967" [ 966.721337] env[61995]: _type = "Task" [ 966.721337] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.731310] env[61995]: DEBUG oslo_vmware.api [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.841609] env[61995]: DEBUG nova.network.neutron [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 966.849536] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794964, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.982497] env[61995]: DEBUG oslo_vmware.api [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794962, 'name': RemoveSnapshot_Task, 'duration_secs': 0.604392} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.982876] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 966.983178] env[61995]: INFO nova.compute.manager [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Took 17.31 seconds to snapshot the instance on the hypervisor. [ 966.991523] env[61995]: DEBUG nova.network.neutron [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updating instance_info_cache with network_info: [{"id": "4554d960-3f34-464e-abd8-8f16166194b3", "address": "fa:16:3e:41:f1:71", "network": {"id": "9dd02d8b-08cb-42e0-a05d-07035ad72737", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1718950085-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978c59760428439cb83c95c9a39be6e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4554d960-3f", "ovs_interfaceid": "4554d960-3f34-464e-abd8-8f16166194b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.004550] env[61995]: DEBUG oslo_vmware.api [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794965, 'name': ReconfigVM_Task, 'duration_secs': 0.169407} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.005472] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 967.149179] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.196357] env[61995]: DEBUG nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.227317] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.227801] env[61995]: DEBUG nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 967.230643] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.350s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.232058] env[61995]: INFO nova.compute.claims [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.241396] env[61995]: DEBUG oslo_vmware.api [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-794967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.308982} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.241641] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.241825] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 967.242012] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 967.242200] env[61995]: INFO nova.compute.manager [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Took 1.16 seconds to destroy the instance on the hypervisor. [ 967.242447] env[61995]: DEBUG oslo.service.loopingcall [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.242642] env[61995]: DEBUG nova.compute.manager [-] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.242737] env[61995]: DEBUG nova.network.neutron [-] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 967.346890] env[61995]: DEBUG oslo_vmware.api [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794964, 'name': PowerOnVM_Task, 'duration_secs': 0.715616} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.347178] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.347388] env[61995]: INFO nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Took 8.91 seconds to spawn the instance on the hypervisor. [ 967.347564] env[61995]: DEBUG nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.348382] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15c67b6-6523-48b9-b06b-2f88a05eacf6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.494013] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Releasing lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.494346] env[61995]: DEBUG nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Instance network_info: |[{"id": "4554d960-3f34-464e-abd8-8f16166194b3", "address": "fa:16:3e:41:f1:71", "network": {"id": "9dd02d8b-08cb-42e0-a05d-07035ad72737", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1718950085-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978c59760428439cb83c95c9a39be6e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4554d960-3f", "ovs_interfaceid": "4554d960-3f34-464e-abd8-8f16166194b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.494786] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:f1:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6966f473-59ac-49bb-9b7a-22c61f4e61e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4554d960-3f34-464e-abd8-8f16166194b3', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.502272] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Creating folder: Project (978c59760428439cb83c95c9a39be6e6). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 967.502851] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7cfee900-cd9a-4ba3-a80f-322b253c95b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.527888] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Created folder: Project (978c59760428439cb83c95c9a39be6e6) in parent group-v185203. [ 967.528149] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Creating folder: Instances. Parent ref: group-v185377. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 967.530676] env[61995]: DEBUG nova.compute.manager [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Found 3 images (rotation: 2) {{(pid=61995) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 967.530872] env[61995]: DEBUG nova.compute.manager [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Rotating out 1 backups {{(pid=61995) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 967.531047] env[61995]: DEBUG nova.compute.manager [None req-41414989-c203-41e9-996a-baa1f32ddb68 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleting image 4c8c4eb7-8495-4e66-b897-ea812737ebc8 {{(pid=61995) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 967.532863] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-196a4fd7-a741-4938-8914-3e54147c2598 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.545147] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Created folder: Instances in parent group-v185377. [ 967.545415] env[61995]: DEBUG oslo.service.loopingcall [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.547360] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 967.547599] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b11b5407-1664-42b4-9ab9-1d9bd209a4c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.563591] env[61995]: DEBUG nova.objects.instance [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'flavor' on Instance uuid bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.572035] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.572035] env[61995]: value = "task-794970" [ 967.572035] env[61995]: _type = "Task" [ 967.572035] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.581900] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794970, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.723521] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.740059] env[61995]: DEBUG nova.compute.utils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.741352] env[61995]: DEBUG nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 967.741490] env[61995]: DEBUG nova.network.neutron [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 967.788421] env[61995]: DEBUG nova.policy [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f27aa509c254e2fa6a5d833b53f50e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4042c97b25a24bf0a1ebe23e958f95ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 967.800126] env[61995]: DEBUG nova.compute.manager [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Received event network-changed-4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.800206] env[61995]: DEBUG nova.compute.manager [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Refreshing instance network info cache due to event network-changed-4554d960-3f34-464e-abd8-8f16166194b3. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 967.800473] env[61995]: DEBUG oslo_concurrency.lockutils [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] Acquiring lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.800643] env[61995]: DEBUG oslo_concurrency.lockutils [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] Acquired lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.800891] env[61995]: DEBUG nova.network.neutron [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Refreshing network info cache for port 4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 967.868182] env[61995]: INFO nova.compute.manager [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Took 18.71 seconds to build instance. [ 967.961867] env[61995]: DEBUG nova.network.neutron [-] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.069524] env[61995]: DEBUG oslo_concurrency.lockutils [None req-023836c7-faa0-4ad3-854b-7854a1d5917d tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 6.846s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.084574] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794970, 'name': CreateVM_Task, 'duration_secs': 0.374522} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.084574] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 968.085205] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.085443] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.085780] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.086098] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71ccb103-4807-4118-b3f8-92a87b5d1b00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.092796] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 968.092796] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a672ba-c32a-3852-2e23-685d85ea2fe9" [ 968.092796] env[61995]: _type = "Task" [ 968.092796] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.110772] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a672ba-c32a-3852-2e23-685d85ea2fe9, 'name': SearchDatastore_Task, 'duration_secs': 0.012368} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.111737] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.112062] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.112388] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.112825] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.113104] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.113953] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5af0eadf-45e8-44dc-bd26-73a72a8ef9cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.125205] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.126112] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 968.126797] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b831e03-3639-4b07-a9b4-31634c488908 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.133457] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 968.133457] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525e4c73-5edc-9cc9-2355-7aea0f01c4d2" [ 968.133457] env[61995]: _type = "Task" [ 968.133457] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.144610] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525e4c73-5edc-9cc9-2355-7aea0f01c4d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.245313] env[61995]: DEBUG nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 968.272792] env[61995]: DEBUG nova.network.neutron [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Successfully created port: 592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 968.371322] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27d399c4-e7b7-4210-9f35-f9189b0144aa tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.220s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.467061] env[61995]: INFO nova.compute.manager [-] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Took 1.22 seconds to deallocate network for instance. [ 968.526130] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66eb375-1be6-411d-8c4e-b6c678929a01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.535708] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b3fa5f-2f18-4abf-8571-b72768546eb8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.568720] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418ca53f-eb45-4439-9c34-faf1b17aa201 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.577095] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be066fb-e5be-4976-a45a-eb6b8e8b0120 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.592889] env[61995]: DEBUG nova.compute.provider_tree [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.604517] env[61995]: INFO nova.compute.manager [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Rescuing [ 968.604787] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.604936] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.605123] env[61995]: DEBUG nova.network.neutron [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 968.644208] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525e4c73-5edc-9cc9-2355-7aea0f01c4d2, 'name': SearchDatastore_Task, 'duration_secs': 0.010921} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.654081] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05234440-b61e-470b-b212-7792448ab6ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.659052] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 968.659052] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a48d63-4d83-bb51-aeb5-6f38bba72fa5" [ 968.659052] env[61995]: _type = "Task" [ 968.659052] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.668210] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a48d63-4d83-bb51-aeb5-6f38bba72fa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.773462] env[61995]: DEBUG nova.network.neutron [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updated VIF entry in instance network info cache for port 4554d960-3f34-464e-abd8-8f16166194b3. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 968.773893] env[61995]: DEBUG nova.network.neutron [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updating instance_info_cache with network_info: [{"id": "4554d960-3f34-464e-abd8-8f16166194b3", "address": "fa:16:3e:41:f1:71", "network": {"id": "9dd02d8b-08cb-42e0-a05d-07035ad72737", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1718950085-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978c59760428439cb83c95c9a39be6e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4554d960-3f", "ovs_interfaceid": "4554d960-3f34-464e-abd8-8f16166194b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.978260] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.096560] env[61995]: DEBUG nova.scheduler.client.report [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.159124] env[61995]: INFO nova.compute.manager [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Rebuilding instance [ 969.171938] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a48d63-4d83-bb51-aeb5-6f38bba72fa5, 'name': SearchDatastore_Task, 'duration_secs': 0.039466} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.172280] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.172577] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 847df13f-8451-49a1-9c6a-7c7c26952940/847df13f-8451-49a1-9c6a-7c7c26952940.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 969.173311] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62af845e-16d0-4420-a2eb-b7e025e35fe6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.182186] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 969.182186] env[61995]: value = "task-794971" [ 969.182186] env[61995]: _type = "Task" [ 969.182186] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.194708] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794971, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.206812] env[61995]: DEBUG nova.compute.manager [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.207630] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba40a2f1-b869-43a0-93e2-0de56e07b198 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.254896] env[61995]: DEBUG nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 969.276828] env[61995]: DEBUG oslo_concurrency.lockutils [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] Releasing lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.277947] env[61995]: DEBUG nova.compute.manager [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Received event network-vif-deleted-0496010d-b03c-4002-b2d4-372cdbededa9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.277947] env[61995]: INFO nova.compute.manager [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Neutron deleted interface 0496010d-b03c-4002-b2d4-372cdbededa9; detaching it from the instance and deleting it from the info cache [ 969.277947] env[61995]: DEBUG nova.network.neutron [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.283118] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 969.283418] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 969.283608] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.283819] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 969.283981] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.284184] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 969.284433] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 969.284579] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 969.284750] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 969.284918] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 969.285117] env[61995]: DEBUG nova.virt.hardware [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.286362] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff3535f-73da-4ce5-b478-360e53e6ca20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.300756] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4fd048-5f25-4abc-b16f-d5008d56bd4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.465416] env[61995]: DEBUG nova.network.neutron [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Updating instance_info_cache with network_info: [{"id": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "address": "fa:16:3e:58:1b:ff", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8ff3c80-bf", "ovs_interfaceid": "d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.602214] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.602835] env[61995]: DEBUG nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 969.606630] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.665s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.606902] env[61995]: DEBUG nova.objects.instance [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lazy-loading 'resources' on Instance uuid 7d987006-afac-4973-ae69-f15b4752469b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.697840] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794971, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.719648] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.720032] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c8bb1a2-d0b4-49f3-ab51-6ac6938ff244 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.733353] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 969.733353] env[61995]: value = "task-794972" [ 969.733353] env[61995]: _type = "Task" [ 969.733353] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.746981] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794972, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.780803] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8785ebc8-7adc-4171-af4b-a6b5b615d3ec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.794289] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5048f2bc-c2fa-4d56-b841-bd09dfb9ef2c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.839563] env[61995]: DEBUG nova.compute.manager [req-caae84e8-c18e-4aad-9fb5-5c1d279aabae req-0080f35f-16bb-42a6-a4ad-09e30132bbe2 service nova] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Detach interface failed, port_id=0496010d-b03c-4002-b2d4-372cdbededa9, reason: Instance 057ba9d6-4ff2-4da1-be7d-ecfadf738d74 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 969.968536] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-5cb4c1d9-e661-48b4-8600-1b823d63ddbb" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.077446] env[61995]: DEBUG nova.compute.manager [req-eb7b780e-b9ac-43a5-bf99-31a47503455e req-75560a26-b7aa-48ff-a4f0-aac567d66b78 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Received event network-vif-plugged-592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.077446] env[61995]: DEBUG oslo_concurrency.lockutils [req-eb7b780e-b9ac-43a5-bf99-31a47503455e req-75560a26-b7aa-48ff-a4f0-aac567d66b78 service nova] Acquiring lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.077446] env[61995]: DEBUG oslo_concurrency.lockutils [req-eb7b780e-b9ac-43a5-bf99-31a47503455e req-75560a26-b7aa-48ff-a4f0-aac567d66b78 service nova] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.077739] env[61995]: DEBUG oslo_concurrency.lockutils [req-eb7b780e-b9ac-43a5-bf99-31a47503455e req-75560a26-b7aa-48ff-a4f0-aac567d66b78 service nova] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.078227] env[61995]: DEBUG nova.compute.manager [req-eb7b780e-b9ac-43a5-bf99-31a47503455e req-75560a26-b7aa-48ff-a4f0-aac567d66b78 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] No waiting events found dispatching network-vif-plugged-592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.078734] env[61995]: WARNING nova.compute.manager [req-eb7b780e-b9ac-43a5-bf99-31a47503455e req-75560a26-b7aa-48ff-a4f0-aac567d66b78 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Received unexpected event network-vif-plugged-592fb7d9-e1e2-4233-80f7-6a263f59fac2 for instance with vm_state building and task_state spawning. [ 970.113028] env[61995]: DEBUG nova.compute.utils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.117027] env[61995]: DEBUG nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.117027] env[61995]: DEBUG nova.network.neutron [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 970.200854] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794971, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.703527} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.203649] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 847df13f-8451-49a1-9c6a-7c7c26952940/847df13f-8451-49a1-9c6a-7c7c26952940.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 970.204148] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.204753] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83025527-cc82-43b3-9bdc-1fb23f508b08 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.215020] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 970.215020] env[61995]: value = "task-794973" [ 970.215020] env[61995]: _type = "Task" [ 970.215020] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.228317] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.247883] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794972, 'name': PowerOffVM_Task, 'duration_secs': 0.248574} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.247883] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 970.312805] env[61995]: INFO nova.compute.manager [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Detaching volume caeb58b4-c8a5-46b3-9f21-57540bd5b068 [ 970.361375] env[61995]: INFO nova.virt.block_device [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Attempting to driver detach volume caeb58b4-c8a5-46b3-9f21-57540bd5b068 from mountpoint /dev/sdb [ 970.361375] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 970.361375] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 970.361375] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4a8a99-1775-443f-a99e-7362462b6d43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.389381] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147b5af7-1f83-4280-a920-5180dc4518cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.410449] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fdb44e-ccbb-4d2e-b4dd-6f3ea9a7e6fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.418781] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31508453-f1f3-45e2-a255-f26b36cd25de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.456564] env[61995]: DEBUG nova.policy [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caf9bbbd848d4d7ba53ba3bd553943e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2018aae65cf4bf98b5ac82519201cf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 970.458406] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe07d8e1-8491-47d1-992c-a79b28f8f9fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.463046] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ee4a1f-e481-403d-941e-005bfea1f71b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.470724] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "9627f278-73d6-45d3-b60d-776e373eef73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.470724] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "9627f278-73d6-45d3-b60d-776e373eef73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.482618] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] The volume has not been displaced from its original location: [datastore2] volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068/volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 970.487922] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 970.513750] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96c3a24f-bd15-4d2a-8b39-4c290f25e8e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.527391] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "229d966c-b70d-4a51-a176-2e88488d5c4b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.527391] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.527391] env[61995]: DEBUG nova.compute.manager [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.530073] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 970.531044] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b52fe75-1795-49b7-9716-f64b671e005d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.534119] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdf2559-877b-4218-9a19-bd7019e22766 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.536498] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1d8ebdf-a9e6-4021-8bb6-16a85865df5b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.538311] env[61995]: DEBUG nova.network.neutron [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Successfully updated port: 592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.547033] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447e2cc8-1fad-4198-bae8-0784a5d4047b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.551627] env[61995]: DEBUG nova.compute.manager [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 970.552346] env[61995]: DEBUG nova.objects.instance [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'flavor' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.556836] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 970.556836] env[61995]: value = "task-794974" [ 970.556836] env[61995]: _type = "Task" [ 970.556836] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.557118] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 970.557118] env[61995]: value = "task-794975" [ 970.557118] env[61995]: _type = "Task" [ 970.557118] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.572374] env[61995]: DEBUG nova.compute.provider_tree [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.583300] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794974, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.583559] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794975, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.619469] env[61995]: DEBUG nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.723787] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072334} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.725087] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.725087] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db91525-cb63-4efa-83da-c192a15f5894 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.747379] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 847df13f-8451-49a1-9c6a-7c7c26952940/847df13f-8451-49a1-9c6a-7c7c26952940.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.747803] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e66551c4-a69e-4fcf-a8c4-3e93318817dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.769659] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 970.769659] env[61995]: value = "task-794976" [ 970.769659] env[61995]: _type = "Task" [ 970.769659] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.779279] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.846945] env[61995]: DEBUG nova.network.neutron [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Successfully created port: 0e1358a1-ed1a-453b-882f-a94d4440e000 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.971635] env[61995]: DEBUG nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.041514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.041514] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.041514] env[61995]: DEBUG nova.network.neutron [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 971.063749] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.066792] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f938b12-1108-47bf-af98-1ffeaced3801 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.074624] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794975, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.079033] env[61995]: DEBUG nova.scheduler.client.report [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.083406] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794974, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.083406] env[61995]: DEBUG oslo_vmware.api [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 971.083406] env[61995]: value = "task-794977" [ 971.083406] env[61995]: _type = "Task" [ 971.083406] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.091514] env[61995]: DEBUG oslo_vmware.api [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794977, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.279378] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.497716] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.573618] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794975, 'name': ReconfigVM_Task, 'duration_secs': 1.008565} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.577082] env[61995]: DEBUG nova.network.neutron [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 971.579227] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 971.583932] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794974, 'name': PowerOffVM_Task, 'duration_secs': 0.658677} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.584844] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.978s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.588231] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1fc2068-017e-4dac-ae81-283e72a278e5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.598914] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.599887] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.451s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.600286] env[61995]: DEBUG nova.objects.instance [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lazy-loading 'resources' on Instance uuid 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.607018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21853263-5951-462d-b47e-32da927760b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.614289] env[61995]: DEBUG oslo_vmware.api [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-794977, 'name': PowerOffVM_Task, 'duration_secs': 0.447632} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.631964] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.632374] env[61995]: DEBUG nova.compute.manager [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.632828] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 971.632828] env[61995]: value = "task-794978" [ 971.632828] env[61995]: _type = "Task" [ 971.632828] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.634260] env[61995]: DEBUG nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.636946] env[61995]: INFO nova.scheduler.client.report [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Deleted allocations for instance 7d987006-afac-4973-ae69-f15b4752469b [ 971.638663] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ec4bba-201d-43c9-979d-9f56c44a47a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.642505] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6083ca3d-11d1-4e18-91e7-d47bf98bb1af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.666068] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794978, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.668203] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.668570] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.668635] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.668786] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.668932] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.669102] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.669375] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.669553] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.669725] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.669889] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.670160] env[61995]: DEBUG nova.virt.hardware [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.670999] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef0658b-e364-48ba-93c9-fea119997f49 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.686281] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.686281] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4ef86d-ca25-4889-97f9-cae7bffa0192 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.688448] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92fdbb59-0668-482e-a9af-65a1867f1ce0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.706430] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 971.706430] env[61995]: value = "task-794979" [ 971.706430] env[61995]: _type = "Task" [ 971.706430] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.719236] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 971.719494] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.719742] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.719886] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.720076] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.720324] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8ea4a63-057e-4aec-aeb2-60909092c392 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.732407] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.732653] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 971.733519] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65670662-95e2-4191-8b91-3ef39dabe8a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.740891] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 971.740891] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52288911-780c-fb49-b399-39be7370740b" [ 971.740891] env[61995]: _type = "Task" [ 971.740891] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.752934] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52288911-780c-fb49-b399-39be7370740b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.779173] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794976, 'name': ReconfigVM_Task, 'duration_secs': 0.833558} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.779509] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 847df13f-8451-49a1-9c6a-7c7c26952940/847df13f-8451-49a1-9c6a-7c7c26952940.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.780132] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-47fb4783-9aac-44e7-9daa-d02d269e271a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.786967] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 971.786967] env[61995]: value = "task-794980" [ 971.786967] env[61995]: _type = "Task" [ 971.786967] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.795809] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794980, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.841442] env[61995]: DEBUG nova.network.neutron [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Updating instance_info_cache with network_info: [{"id": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "address": "fa:16:3e:79:c1:cd", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap592fb7d9-e1", "ovs_interfaceid": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.109083] env[61995]: DEBUG nova.compute.manager [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Received event network-changed-592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.109384] env[61995]: DEBUG nova.compute.manager [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Refreshing instance network info cache due to event network-changed-592fb7d9-e1e2-4233-80f7-6a263f59fac2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 972.109526] env[61995]: DEBUG oslo_concurrency.lockutils [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] Acquiring lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.155659] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794978, 'name': ReconfigVM_Task, 'duration_secs': 0.151934} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.156126] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0061946e-c9e4-4546-bfc0-8433c5fc6d45 tempest-ServersTestBootFromVolume-1142411735 tempest-ServersTestBootFromVolume-1142411735-project-member] Lock "7d987006-afac-4973-ae69-f15b4752469b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.194s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.159229] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 972.168929] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5764659e-ce16-44d7-badf-bef027cd2a94 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.642s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.260635] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52288911-780c-fb49-b399-39be7370740b, 'name': SearchDatastore_Task, 'duration_secs': 0.017906} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.261661] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45239bf0-89bf-4760-be17-4fbfd8d7b8bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.267961] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 972.267961] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5290d4a6-1c6d-bd13-5e0b-9d1a4c11ae10" [ 972.267961] env[61995]: _type = "Task" [ 972.267961] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.278182] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5290d4a6-1c6d-bd13-5e0b-9d1a4c11ae10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.299234] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794980, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.346256] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.346638] env[61995]: DEBUG nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Instance network_info: |[{"id": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "address": "fa:16:3e:79:c1:cd", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap592fb7d9-e1", "ovs_interfaceid": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 972.346969] env[61995]: DEBUG oslo_concurrency.lockutils [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] Acquired lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.347171] env[61995]: DEBUG nova.network.neutron [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Refreshing network info cache for port 592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 972.348520] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:c1:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3291573-fad8-48cc-a965-c3554e7cee4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '592fb7d9-e1e2-4233-80f7-6a263f59fac2', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.356207] env[61995]: DEBUG oslo.service.loopingcall [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.361397] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 972.362108] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b711316-84af-454a-a2b3-537b61c70a8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.380161] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91ff5a9-3caa-4ec5-b53a-a5fd2332ed02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.393023] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f02ba1f-680d-4515-8057-8917f3cd8794 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.395704] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.395704] env[61995]: value = "task-794981" [ 972.395704] env[61995]: _type = "Task" [ 972.395704] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.429666] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c56780f-105a-43a5-9dac-65be56b7c645 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.436433] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794981, 'name': CreateVM_Task} progress is 15%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.442106] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb56b87-b091-4cd9-a6c9-0e25c993896b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.456418] env[61995]: DEBUG nova.compute.provider_tree [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.520641] env[61995]: DEBUG nova.network.neutron [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Successfully updated port: 0e1358a1-ed1a-453b-882f-a94d4440e000 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.781635] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5290d4a6-1c6d-bd13-5e0b-9d1a4c11ae10, 'name': SearchDatastore_Task, 'duration_secs': 0.027898} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.781635] env[61995]: DEBUG oslo_concurrency.lockutils [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.781876] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk. {{(pid=61995) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 972.782068] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d235255f-9e57-4bb8-abca-82425ee2ecf0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.792913] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 972.792913] env[61995]: value = "task-794982" [ 972.792913] env[61995]: _type = "Task" [ 972.792913] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.800378] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794980, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.806112] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794982, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.807053] env[61995]: DEBUG nova.network.neutron [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Updated VIF entry in instance network info cache for port 592fb7d9-e1e2-4233-80f7-6a263f59fac2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 972.807404] env[61995]: DEBUG nova.network.neutron [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Updating instance_info_cache with network_info: [{"id": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "address": "fa:16:3e:79:c1:cd", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap592fb7d9-e1", "ovs_interfaceid": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.906915] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794981, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.959897] env[61995]: DEBUG nova.scheduler.client.report [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.025040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "refresh_cache-8ae65d63-16c7-4c67-a8c2-73968f63457d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.025040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "refresh_cache-8ae65d63-16c7-4c67-a8c2-73968f63457d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.025040] env[61995]: DEBUG nova.network.neutron [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 973.221651] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 973.221651] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea1969e7-5463-4678-a66b-1a7ca881d5ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.232565] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 973.232565] env[61995]: value = "task-794983" [ 973.232565] env[61995]: _type = "Task" [ 973.232565] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.247383] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 973.247383] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 973.247383] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 973.252029] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d50dc7c-1bda-4868-8fb4-9d29c70376ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.272582] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3455574-8146-4e58-a070-4604de2d7a59 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.283241] env[61995]: WARNING nova.virt.vmwareapi.driver [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 973.283547] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 973.284461] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6b0fe5-532b-48a7-b70b-26bb7647da37 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.298174] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 973.302582] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b312e0d0-71f3-47ca-873e-284115b50c99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.309140] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794980, 'name': Rename_Task, 'duration_secs': 1.200974} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.309740] env[61995]: DEBUG oslo_concurrency.lockutils [req-fcc36910-d720-49ca-8ff1-4361f562009a req-b8913e09-1593-4cc9-9f4b-698c40e46d76 service nova] Releasing lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.310183] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 973.310433] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10b8756c-701e-4766-8a2a-3ddbd4627407 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.315936] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794982, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.323822] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 973.323822] env[61995]: value = "task-794985" [ 973.323822] env[61995]: _type = "Task" [ 973.323822] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.335832] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.411051] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794981, 'name': CreateVM_Task, 'duration_secs': 0.826391} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.411298] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 973.412060] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.412244] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.412604] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 973.412939] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df180930-fdc2-422a-91a8-adf38f3f4d62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.420380] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 973.420380] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a2d089-8d9a-18e8-52db-8cff437d88fe" [ 973.420380] env[61995]: _type = "Task" [ 973.420380] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.430289] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a2d089-8d9a-18e8-52db-8cff437d88fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.466788] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.867s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.469625] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.746s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.471381] env[61995]: INFO nova.compute.claims [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 973.487146] env[61995]: INFO nova.scheduler.client.report [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Deleted allocations for instance 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5 [ 973.578097] env[61995]: DEBUG nova.network.neutron [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 973.803344] env[61995]: DEBUG nova.network.neutron [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Updating instance_info_cache with network_info: [{"id": "0e1358a1-ed1a-453b-882f-a94d4440e000", "address": "fa:16:3e:c1:a0:4c", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e1358a1-ed", "ovs_interfaceid": "0e1358a1-ed1a-453b-882f-a94d4440e000", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.811374] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794982, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.703879} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.811719] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk. [ 973.812415] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9563f2d7-3a75-40bc-8570-53fc8eddef57 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.839740] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.843489] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8fd73e2-51d9-459f-89fb-e34eaab610db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.863522] env[61995]: DEBUG oslo_vmware.api [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-794985, 'name': PowerOnVM_Task, 'duration_secs': 0.518982} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.864792] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 973.865015] env[61995]: INFO nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Took 9.34 seconds to spawn the instance on the hypervisor. [ 973.865218] env[61995]: DEBUG nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.865547] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 973.865547] env[61995]: value = "task-794986" [ 973.865547] env[61995]: _type = "Task" [ 973.865547] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.866285] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87975ec8-03cf-4803-ab54-3f9ae5147d2c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.876458] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794986, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.931866] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a2d089-8d9a-18e8-52db-8cff437d88fe, 'name': SearchDatastore_Task, 'duration_secs': 0.020884} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.932166] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.933195] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 973.933195] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.933195] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.933195] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 973.933506] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f4b6121-ea92-44fb-9fee-1a484030a9f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.942989] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 973.943191] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 973.943927] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fae46416-f201-4365-ac2d-3f098d9881eb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.952270] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 973.952270] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5276848b-bbb7-b130-0a17-4dcc2b98bfaf" [ 973.952270] env[61995]: _type = "Task" [ 973.952270] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.960687] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5276848b-bbb7-b130-0a17-4dcc2b98bfaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.994696] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c12b9f0f-30ba-4cbd-8c4e-b5b08e4a44ed tempest-ServersNegativeTestJSON-1087591599 tempest-ServersNegativeTestJSON-1087591599-project-member] Lock "0e152049-f9ca-4ac1-b12d-ae1ad4623ff5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.811s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.073895] env[61995]: DEBUG nova.compute.manager [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Stashing vm_state: stopped {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 974.237726] env[61995]: DEBUG nova.compute.manager [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Received event network-vif-plugged-0e1358a1-ed1a-453b-882f-a94d4440e000 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.237984] env[61995]: DEBUG oslo_concurrency.lockutils [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] Acquiring lock "8ae65d63-16c7-4c67-a8c2-73968f63457d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.238872] env[61995]: DEBUG oslo_concurrency.lockutils [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.239055] env[61995]: DEBUG oslo_concurrency.lockutils [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.239437] env[61995]: DEBUG nova.compute.manager [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] No waiting events found dispatching network-vif-plugged-0e1358a1-ed1a-453b-882f-a94d4440e000 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 974.239663] env[61995]: WARNING nova.compute.manager [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Received unexpected event network-vif-plugged-0e1358a1-ed1a-453b-882f-a94d4440e000 for instance with vm_state building and task_state spawning. [ 974.239842] env[61995]: DEBUG nova.compute.manager [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Received event network-changed-0e1358a1-ed1a-453b-882f-a94d4440e000 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.240018] env[61995]: DEBUG nova.compute.manager [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Refreshing instance network info cache due to event network-changed-0e1358a1-ed1a-453b-882f-a94d4440e000. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 974.240196] env[61995]: DEBUG oslo_concurrency.lockutils [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] Acquiring lock "refresh_cache-8ae65d63-16c7-4c67-a8c2-73968f63457d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.306767] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "refresh_cache-8ae65d63-16c7-4c67-a8c2-73968f63457d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.307261] env[61995]: DEBUG nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Instance network_info: |[{"id": "0e1358a1-ed1a-453b-882f-a94d4440e000", "address": "fa:16:3e:c1:a0:4c", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e1358a1-ed", "ovs_interfaceid": "0e1358a1-ed1a-453b-882f-a94d4440e000", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 974.307820] env[61995]: DEBUG oslo_concurrency.lockutils [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] Acquired lock "refresh_cache-8ae65d63-16c7-4c67-a8c2-73968f63457d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.308050] env[61995]: DEBUG nova.network.neutron [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Refreshing network info cache for port 0e1358a1-ed1a-453b-882f-a94d4440e000 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.309321] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:a0:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e1358a1-ed1a-453b-882f-a94d4440e000', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.324288] env[61995]: DEBUG oslo.service.loopingcall [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.328379] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 974.328929] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-470f9ebe-e26b-4e47-9bbf-27bef93b2dcc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.352087] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.352087] env[61995]: value = "task-794987" [ 974.352087] env[61995]: _type = "Task" [ 974.352087] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.360727] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794987, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.379779] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794986, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.391139] env[61995]: INFO nova.compute.manager [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Took 19.11 seconds to build instance. [ 974.469547] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5276848b-bbb7-b130-0a17-4dcc2b98bfaf, 'name': SearchDatastore_Task, 'duration_secs': 0.015491} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.470507] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2a820ac-3df7-47b3-9a49-c0c20254eddd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.485414] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 974.485414] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ecf458-72b6-74ff-da2c-e12b1bfd1e3f" [ 974.485414] env[61995]: _type = "Task" [ 974.485414] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.498906] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ecf458-72b6-74ff-da2c-e12b1bfd1e3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.609054] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.715351] env[61995]: DEBUG nova.network.neutron [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Updated VIF entry in instance network info cache for port 0e1358a1-ed1a-453b-882f-a94d4440e000. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 974.715792] env[61995]: DEBUG nova.network.neutron [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Updating instance_info_cache with network_info: [{"id": "0e1358a1-ed1a-453b-882f-a94d4440e000", "address": "fa:16:3e:c1:a0:4c", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e1358a1-ed", "ovs_interfaceid": "0e1358a1-ed1a-453b-882f-a94d4440e000", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.767177] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91fba43-ec5d-42a0-a463-bed1b8fe1578 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.775273] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e7cbb8-1d23-4116-b1d2-184b1a6bcb7f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.811430] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2940dcdc-11fc-4f52-8db6-920c994a802b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.821570] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6b8aed-6c03-4e2a-b961-ecb298bf4388 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.836664] env[61995]: DEBUG nova.compute.provider_tree [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.861943] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-794987, 'name': CreateVM_Task, 'duration_secs': 0.476077} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.862300] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 974.863011] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.863238] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.863609] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.863954] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25636195-ee64-461a-92f8-bf0f09a8da25 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.871143] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 974.871143] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5204560d-5aa3-f003-7cfa-ca3a13ee5c82" [ 974.871143] env[61995]: _type = "Task" [ 974.871143] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.883541] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794986, 'name': ReconfigVM_Task, 'duration_secs': 0.807886} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.887441] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.888076] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5204560d-5aa3-f003-7cfa-ca3a13ee5c82, 'name': SearchDatastore_Task, 'duration_secs': 0.009778} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.888812] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576b871a-96d9-42b0-915a-e7f76b288d65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.891475] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.891769] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.892040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.892861] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5bde0bb9-56ad-4a62-9acf-7ddabcbfbfb3 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "847df13f-8451-49a1-9c6a-7c7c26952940" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.640s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.917123] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3933e32f-19c0-4153-b75b-0b5207680d21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.940083] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 974.940083] env[61995]: value = "task-794988" [ 974.940083] env[61995]: _type = "Task" [ 974.940083] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.949217] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794988, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.998297] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ecf458-72b6-74ff-da2c-e12b1bfd1e3f, 'name': SearchDatastore_Task, 'duration_secs': 0.019314} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.998588] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.998857] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] ba89ee63-e293-47e1-90ab-7b8e72dd1b50/ba89ee63-e293-47e1-90ab-7b8e72dd1b50.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 974.999385] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.999385] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.999628] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e369701-84ec-4401-89ff-df045ba78e18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.001718] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42000e21-a028-4fae-884c-11d9617f5c45 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.011600] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 975.011600] env[61995]: value = "task-794989" [ 975.011600] env[61995]: _type = "Task" [ 975.011600] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.012811] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.012991] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 975.016552] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc59ee98-4a38-4952-951d-f1fe1bdc658a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.024436] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 975.024436] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e36a66-059f-719c-d4d3-760934079f77" [ 975.024436] env[61995]: _type = "Task" [ 975.024436] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.027550] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.035636] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e36a66-059f-719c-d4d3-760934079f77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.218987] env[61995]: DEBUG oslo_concurrency.lockutils [req-f78b0900-f088-429f-a20a-5aaf1e191f8b req-8b97f995-1540-45fc-8501-95b35ac11e58 service nova] Releasing lock "refresh_cache-8ae65d63-16c7-4c67-a8c2-73968f63457d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.342728] env[61995]: DEBUG nova.scheduler.client.report [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 975.454134] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794988, 'name': ReconfigVM_Task, 'duration_secs': 0.190714} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.454624] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 975.455131] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bdc6d393-8bdc-46ed-af64-3402e63920cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.463832] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 975.463832] env[61995]: value = "task-794990" [ 975.463832] env[61995]: _type = "Task" [ 975.463832] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.474243] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794990, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.523628] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794989, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490351} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.523909] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] ba89ee63-e293-47e1-90ab-7b8e72dd1b50/ba89ee63-e293-47e1-90ab-7b8e72dd1b50.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 975.524877] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.524877] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82b88e75-7e20-4a98-bcdf-597943100780 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.533464] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 975.533464] env[61995]: value = "task-794991" [ 975.533464] env[61995]: _type = "Task" [ 975.533464] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.540506] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e36a66-059f-719c-d4d3-760934079f77, 'name': SearchDatastore_Task, 'duration_secs': 0.010155} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.541716] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86a6e6e1-1657-400f-9c28-07f72102cc55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.547294] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794991, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.551088] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 975.551088] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525e24f7-aa9c-9f8d-62e1-5d9a64103a1e" [ 975.551088] env[61995]: _type = "Task" [ 975.551088] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.559714] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525e24f7-aa9c-9f8d-62e1-5d9a64103a1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.848901] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.849530] env[61995]: DEBUG nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 975.852201] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.874s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.852981] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.854417] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.357s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.856292] env[61995]: INFO nova.compute.claims [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.912616] env[61995]: INFO nova.scheduler.client.report [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted allocations for instance 057ba9d6-4ff2-4da1-be7d-ecfadf738d74 [ 975.978017] env[61995]: DEBUG oslo_vmware.api [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-794990, 'name': PowerOnVM_Task, 'duration_secs': 0.460479} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.978433] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 975.981487] env[61995]: DEBUG nova.compute.manager [None req-27edbb5f-930f-4584-a4b1-5c02b8da0247 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.982321] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37bd1e81-17ba-407d-a23c-50131ee429b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.046048] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794991, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074245} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.046530] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.050194] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff51afbd-3d9d-4cb3-aeed-0960708a4486 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.075868] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] ba89ee63-e293-47e1-90ab-7b8e72dd1b50/ba89ee63-e293-47e1-90ab-7b8e72dd1b50.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.077013] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd142488-b935-4f12-a707-f72010401fa9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.096325] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525e24f7-aa9c-9f8d-62e1-5d9a64103a1e, 'name': SearchDatastore_Task, 'duration_secs': 0.014937} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.097075] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.097393] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 8ae65d63-16c7-4c67-a8c2-73968f63457d/8ae65d63-16c7-4c67-a8c2-73968f63457d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 976.097751] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e7e6739-3b87-4973-9fd2-3e5ee6349e12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.105152] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 976.105152] env[61995]: value = "task-794992" [ 976.105152] env[61995]: _type = "Task" [ 976.105152] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.108898] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 976.108898] env[61995]: value = "task-794993" [ 976.108898] env[61995]: _type = "Task" [ 976.108898] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.116713] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794992, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.122877] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.363800] env[61995]: DEBUG nova.compute.utils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 976.365970] env[61995]: DEBUG nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 976.366167] env[61995]: DEBUG nova.network.neutron [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 976.415693] env[61995]: DEBUG nova.compute.manager [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Received event network-changed-4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.415693] env[61995]: DEBUG nova.compute.manager [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Refreshing instance network info cache due to event network-changed-4554d960-3f34-464e-abd8-8f16166194b3. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 976.415693] env[61995]: DEBUG oslo_concurrency.lockutils [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] Acquiring lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.415693] env[61995]: DEBUG oslo_concurrency.lockutils [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] Acquired lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.415693] env[61995]: DEBUG nova.network.neutron [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Refreshing network info cache for port 4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 976.421464] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ddc9bc13-ab9c-48fd-8d00-c7a5a911f32f tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "057ba9d6-4ff2-4da1-be7d-ecfadf738d74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.343s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.438636] env[61995]: DEBUG nova.policy [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2388a3bd1587413285ff9f9ec07739b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9301d14211a1464eb740d9bb745b631a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 976.575310] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 976.575310] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 976.575310] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleting the datastore file [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.575310] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1f58d88-cb87-404c-9ef0-f58c92b966ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.585293] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 976.585293] env[61995]: value = "task-794994" [ 976.585293] env[61995]: _type = "Task" [ 976.585293] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.596322] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794994, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.621321] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794992, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.626563] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794993, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.866965] env[61995]: DEBUG nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 977.104251] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-794994, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.390394} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.104535] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.104742] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 977.104933] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 977.131748] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794992, 'name': ReconfigVM_Task, 'duration_secs': 0.518695} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.134771] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Reconfigured VM instance instance-00000052 to attach disk [datastore2] ba89ee63-e293-47e1-90ab-7b8e72dd1b50/ba89ee63-e293-47e1-90ab-7b8e72dd1b50.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.135331] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794993, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.700823} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.136526] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a50cdcb-5152-41d9-be51-ada090be1e12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.137392] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 8ae65d63-16c7-4c67-a8c2-73968f63457d/8ae65d63-16c7-4c67-a8c2-73968f63457d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 977.137485] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.137729] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ca1d751-147c-4925-97d9-ca89d28bf359 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.148617] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 977.148617] env[61995]: value = "task-794995" [ 977.148617] env[61995]: _type = "Task" [ 977.148617] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.148617] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 977.148617] env[61995]: value = "task-794996" [ 977.148617] env[61995]: _type = "Task" [ 977.148617] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.164592] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b54381-13bb-4c93-9041-19c5bfc9ec1b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.176550] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794995, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.181026] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794996, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.182834] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85015107-fda2-4789-a6f2-0da31e0e33c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.222549] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d7dca8-e3a8-4bac-8e9c-33b5649f9054 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.235193] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c8ce59-07a2-44ff-b894-2fb9db8e0189 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.250457] env[61995]: DEBUG nova.compute.provider_tree [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.252426] env[61995]: DEBUG nova.network.neutron [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Successfully created port: fd56d4b0-9fa3-445a-a99c-67842d5494e7 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 977.557150] env[61995]: DEBUG nova.network.neutron [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updated VIF entry in instance network info cache for port 4554d960-3f34-464e-abd8-8f16166194b3. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.557789] env[61995]: DEBUG nova.network.neutron [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updating instance_info_cache with network_info: [{"id": "4554d960-3f34-464e-abd8-8f16166194b3", "address": "fa:16:3e:41:f1:71", "network": {"id": "9dd02d8b-08cb-42e0-a05d-07035ad72737", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1718950085-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978c59760428439cb83c95c9a39be6e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4554d960-3f", "ovs_interfaceid": "4554d960-3f34-464e-abd8-8f16166194b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.619839] env[61995]: INFO nova.virt.block_device [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Booting with volume caeb58b4-c8a5-46b3-9f21-57540bd5b068 at /dev/sdb [ 977.658691] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794995, 'name': Rename_Task, 'duration_secs': 0.444942} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.662296] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 977.663589] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c755c08b-3b22-4d2f-9e0c-72bf4e6aac76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.666432] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8f4da2d-2e0a-4723-a07c-9a503cd18862 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.673684] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 977.673684] env[61995]: value = "task-794997" [ 977.673684] env[61995]: _type = "Task" [ 977.673684] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.677886] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794996, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.34064} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.681508] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.684261] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00da3c14-7d22-4ba8-a130-169b28cacc02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.696982] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777ac5a2-dc36-4989-90a0-e4289fbdece7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.722322] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 8ae65d63-16c7-4c67-a8c2-73968f63457d/8ae65d63-16c7-4c67-a8c2-73968f63457d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.735406] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-908a9dee-9a05-4085-94a4-3574d9ec9674 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.749801] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794997, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.753887] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08dd9b32-ddae-4748-a6b1-6c2055229184 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.755531] env[61995]: DEBUG nova.scheduler.client.report [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.761183] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 977.761183] env[61995]: value = "task-794998" [ 977.761183] env[61995]: _type = "Task" [ 977.761183] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.764416] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77dffce9-1de2-49d8-8535-3db00ee8d98e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.785082] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.801420] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa8289a-2ce0-4b80-b24a-d831704dffac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.808877] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4eca72-d269-416e-aac3-eec91c269bb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.826286] env[61995]: DEBUG nova.virt.block_device [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updating existing volume attachment record: 9d764864-b88f-479c-85d5-b378581dd175 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 977.880142] env[61995]: DEBUG nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 977.911476] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 977.911746] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 977.911908] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.912109] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 977.912274] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.912423] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 977.912618] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 977.912777] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 977.912942] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 977.913154] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 977.913339] env[61995]: DEBUG nova.virt.hardware [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 977.914227] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2a8254-0c8e-48c9-8e4d-6b1bbc74209c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.923429] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5780203a-c57e-4dd8-9b2c-881b7f52bc53 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.060945] env[61995]: DEBUG oslo_concurrency.lockutils [req-dcf6db7d-23d7-4c79-ab15-8f3eed633a96 req-0c6f5f70-a83d-4d54-ba2e-6523a92e3a8a service nova] Releasing lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.191085] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794997, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.263371] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.263951] env[61995]: DEBUG nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.266597] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.658s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.286357] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794998, 'name': ReconfigVM_Task, 'duration_secs': 0.397906} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.289723] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 8ae65d63-16c7-4c67-a8c2-73968f63457d/8ae65d63-16c7-4c67-a8c2-73968f63457d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.290608] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2fc1109-ecfe-4800-a845-9be3a39e6f13 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.298670] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 978.298670] env[61995]: value = "task-794999" [ 978.298670] env[61995]: _type = "Task" [ 978.298670] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.312733] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794999, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.693467] env[61995]: DEBUG oslo_vmware.api [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-794997, 'name': PowerOnVM_Task, 'duration_secs': 0.747988} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.693467] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 978.693467] env[61995]: INFO nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Took 9.44 seconds to spawn the instance on the hypervisor. [ 978.693467] env[61995]: DEBUG nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.694951] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f24b91-c913-40ab-8198-84d8b5ca1e86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.769539] env[61995]: DEBUG nova.compute.utils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.770995] env[61995]: DEBUG nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.771175] env[61995]: DEBUG nova.network.neutron [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 978.775504] env[61995]: INFO nova.compute.claims [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.794445] env[61995]: DEBUG nova.compute.manager [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Received event network-changed-4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.794640] env[61995]: DEBUG nova.compute.manager [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Refreshing instance network info cache due to event network-changed-4554d960-3f34-464e-abd8-8f16166194b3. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 978.794854] env[61995]: DEBUG oslo_concurrency.lockutils [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] Acquiring lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.795352] env[61995]: DEBUG oslo_concurrency.lockutils [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] Acquired lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.795352] env[61995]: DEBUG nova.network.neutron [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Refreshing network info cache for port 4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.812635] env[61995]: INFO nova.compute.manager [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Rescuing [ 978.812886] env[61995]: DEBUG oslo_concurrency.lockutils [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.813055] env[61995]: DEBUG oslo_concurrency.lockutils [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.813228] env[61995]: DEBUG nova.network.neutron [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.818557] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794999, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.867561] env[61995]: DEBUG nova.policy [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e08719d59b456786fce984cc1508e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c3e03b3825c43ca987d60b339fc0a52', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 979.035610] env[61995]: DEBUG nova.compute.manager [req-7d1c9c06-e510-45c5-a7d5-c1a46c6501c7 req-9a52d744-1043-4ed2-b4fc-8e7743d81870 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Received event network-vif-plugged-fd56d4b0-9fa3-445a-a99c-67842d5494e7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 979.035838] env[61995]: DEBUG oslo_concurrency.lockutils [req-7d1c9c06-e510-45c5-a7d5-c1a46c6501c7 req-9a52d744-1043-4ed2-b4fc-8e7743d81870 service nova] Acquiring lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.036091] env[61995]: DEBUG oslo_concurrency.lockutils [req-7d1c9c06-e510-45c5-a7d5-c1a46c6501c7 req-9a52d744-1043-4ed2-b4fc-8e7743d81870 service nova] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.036272] env[61995]: DEBUG oslo_concurrency.lockutils [req-7d1c9c06-e510-45c5-a7d5-c1a46c6501c7 req-9a52d744-1043-4ed2-b4fc-8e7743d81870 service nova] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.036444] env[61995]: DEBUG nova.compute.manager [req-7d1c9c06-e510-45c5-a7d5-c1a46c6501c7 req-9a52d744-1043-4ed2-b4fc-8e7743d81870 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] No waiting events found dispatching network-vif-plugged-fd56d4b0-9fa3-445a-a99c-67842d5494e7 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 979.036612] env[61995]: WARNING nova.compute.manager [req-7d1c9c06-e510-45c5-a7d5-c1a46c6501c7 req-9a52d744-1043-4ed2-b4fc-8e7743d81870 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Received unexpected event network-vif-plugged-fd56d4b0-9fa3-445a-a99c-67842d5494e7 for instance with vm_state building and task_state spawning. [ 979.181563] env[61995]: DEBUG nova.network.neutron [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Successfully updated port: fd56d4b0-9fa3-445a-a99c-67842d5494e7 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.223133] env[61995]: INFO nova.compute.manager [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Took 22.54 seconds to build instance. [ 979.274340] env[61995]: DEBUG nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 979.285862] env[61995]: INFO nova.compute.resource_tracker [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating resource usage from migration 82d397d5-8c3f-445a-ab18-e97956239d3f [ 979.316496] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794999, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.442120] env[61995]: DEBUG nova.network.neutron [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Successfully created port: 0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.558914] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d901702-1e95-45f8-be35-1f4144816a6c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.567287] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9afbe5-b403-49d9-8d99-fce692d56905 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.606270] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643faa32-1543-4f1f-b969-3abf94caedfc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.615554] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9315c3-7225-4e62-9623-f4c734f8a722 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.630380] env[61995]: DEBUG nova.compute.provider_tree [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.666386] env[61995]: DEBUG nova.network.neutron [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updating instance_info_cache with network_info: [{"id": "2327f433-5c63-4d83-8050-6661d9ad60eb", "address": "fa:16:3e:d0:20:ad", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2327f433-5c", "ovs_interfaceid": "2327f433-5c63-4d83-8050-6661d9ad60eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.683873] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "refresh_cache-3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.684298] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "refresh_cache-3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.684526] env[61995]: DEBUG nova.network.neutron [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 979.727539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3fb83418-4ff2-4746-82e7-7c1464307186 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.057s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.809616] env[61995]: DEBUG nova.network.neutron [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updated VIF entry in instance network info cache for port 4554d960-3f34-464e-abd8-8f16166194b3. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.809953] env[61995]: DEBUG nova.network.neutron [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updating instance_info_cache with network_info: [{"id": "4554d960-3f34-464e-abd8-8f16166194b3", "address": "fa:16:3e:41:f1:71", "network": {"id": "9dd02d8b-08cb-42e0-a05d-07035ad72737", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1718950085-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978c59760428439cb83c95c9a39be6e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4554d960-3f", "ovs_interfaceid": "4554d960-3f34-464e-abd8-8f16166194b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.818694] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-794999, 'name': Rename_Task, 'duration_secs': 1.510906} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.819065] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 979.819367] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad12bca1-112b-4e53-afb3-2397ee3ba94e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.827622] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 979.827622] env[61995]: value = "task-795000" [ 979.827622] env[61995]: _type = "Task" [ 979.827622] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.836627] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.970009] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.970295] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.970466] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.970656] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.970809] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.970962] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.971440] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.971686] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.971884] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.972075] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.972254] env[61995]: DEBUG nova.virt.hardware [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.973310] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50efd0f0-423d-4a65-8914-2a42e2435177 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.983982] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49664c6a-36e2-419c-9a8d-b2ca5fca1894 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.000615] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:4f:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0291e27-4d18-41af-ba79-3d058998150e', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.008644] env[61995]: DEBUG oslo.service.loopingcall [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.008971] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 980.009486] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35833aa2-1383-4ad4-86f3-0fd633055869 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.032646] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.032646] env[61995]: value = "task-795001" [ 980.032646] env[61995]: _type = "Task" [ 980.032646] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.042931] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795001, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.136811] env[61995]: DEBUG nova.scheduler.client.report [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.171285] env[61995]: DEBUG oslo_concurrency.lockutils [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.260869] env[61995]: DEBUG nova.network.neutron [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.284210] env[61995]: DEBUG nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.316643] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.317044] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.317312] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.317602] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.318483] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.318483] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.318483] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.318962] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.318962] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.319197] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.319562] env[61995]: DEBUG nova.virt.hardware [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.320436] env[61995]: DEBUG oslo_concurrency.lockutils [req-bb7700ee-6087-4c4c-933d-78823ca8123f req-6918966e-1b27-44aa-a70d-2199289bbe51 service nova] Releasing lock "refresh_cache-847df13f-8451-49a1-9c6a-7c7c26952940" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.321835] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6ed036-ca3c-47b5-8295-7ba7ef9156fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.346078] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396934e1-6558-49be-bc68-9cb8f3245f99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.366111] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795000, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.544867] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795001, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.625865] env[61995]: DEBUG nova.network.neutron [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Updating instance_info_cache with network_info: [{"id": "fd56d4b0-9fa3-445a-a99c-67842d5494e7", "address": "fa:16:3e:d9:01:4b", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd56d4b0-9f", "ovs_interfaceid": "fd56d4b0-9fa3-445a-a99c-67842d5494e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.645440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.379s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.645735] env[61995]: INFO nova.compute.manager [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Migrating [ 980.700761] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 980.701195] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84c2a016-3273-4651-aa51-64a704daecc6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.710128] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 980.710128] env[61995]: value = "task-795002" [ 980.710128] env[61995]: _type = "Task" [ 980.710128] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.721462] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.824076] env[61995]: DEBUG nova.compute.manager [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Received event network-changed-fd56d4b0-9fa3-445a-a99c-67842d5494e7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.824368] env[61995]: DEBUG nova.compute.manager [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Refreshing instance network info cache due to event network-changed-fd56d4b0-9fa3-445a-a99c-67842d5494e7. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 980.824498] env[61995]: DEBUG oslo_concurrency.lockutils [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] Acquiring lock "refresh_cache-3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.839157] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "847df13f-8451-49a1-9c6a-7c7c26952940" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.840539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "847df13f-8451-49a1-9c6a-7c7c26952940" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.840539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "847df13f-8451-49a1-9c6a-7c7c26952940-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.840539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "847df13f-8451-49a1-9c6a-7c7c26952940-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.840539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "847df13f-8451-49a1-9c6a-7c7c26952940-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.842772] env[61995]: INFO nova.compute.manager [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Terminating instance [ 980.848537] env[61995]: DEBUG oslo_vmware.api [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795000, 'name': PowerOnVM_Task, 'duration_secs': 0.55162} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.849157] env[61995]: DEBUG nova.compute.manager [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.849353] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 980.849638] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 980.849827] env[61995]: INFO nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Took 9.22 seconds to spawn the instance on the hypervisor. [ 980.849996] env[61995]: DEBUG nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.850794] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1914d3-8118-4059-ac76-24671f1bad46 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.854858] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b583d58-f934-4de4-b382-d1b380881ca7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.864119] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 980.866891] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-324726e3-67dc-44c8-bb74-e0d4e57e93ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.880215] env[61995]: DEBUG oslo_vmware.api [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 980.880215] env[61995]: value = "task-795003" [ 980.880215] env[61995]: _type = "Task" [ 980.880215] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.891046] env[61995]: DEBUG oslo_vmware.api [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-795003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.046858] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795001, 'name': CreateVM_Task, 'duration_secs': 0.667628} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.047043] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 981.047858] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.048145] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.048568] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 981.048899] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-764b9918-7733-462a-89b0-fba0e60fbbe4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.055186] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 981.055186] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52365df3-5f67-dc38-9344-b175ca5422fe" [ 981.055186] env[61995]: _type = "Task" [ 981.055186] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.065498] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52365df3-5f67-dc38-9344-b175ca5422fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.129315] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "refresh_cache-3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.129753] env[61995]: DEBUG nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Instance network_info: |[{"id": "fd56d4b0-9fa3-445a-a99c-67842d5494e7", "address": "fa:16:3e:d9:01:4b", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd56d4b0-9f", "ovs_interfaceid": "fd56d4b0-9fa3-445a-a99c-67842d5494e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 981.130181] env[61995]: DEBUG oslo_concurrency.lockutils [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] Acquired lock "refresh_cache-3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.130511] env[61995]: DEBUG nova.network.neutron [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Refreshing network info cache for port fd56d4b0-9fa3-445a-a99c-67842d5494e7 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.132718] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:01:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd56d4b0-9fa3-445a-a99c-67842d5494e7', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.143747] env[61995]: DEBUG oslo.service.loopingcall [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.144955] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 981.145243] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b34f797c-f9d4-4deb-b92e-e0c8e639dbca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.166548] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.166748] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.166929] env[61995]: DEBUG nova.network.neutron [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 981.171696] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.171696] env[61995]: value = "task-795004" [ 981.171696] env[61995]: _type = "Task" [ 981.171696] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.182925] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795004, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.222230] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795002, 'name': PowerOffVM_Task, 'duration_secs': 0.312015} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.222597] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 981.223495] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcb1676-4c23-4838-880b-4e6c40b9f35c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.246131] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780b7871-0e9e-4d5b-9c71-bcd7163200d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.301931] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 981.302267] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e49641bb-bb21-4f6a-b014-60e80015a179 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.314892] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 981.314892] env[61995]: value = "task-795005" [ 981.314892] env[61995]: _type = "Task" [ 981.314892] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.324363] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.385625] env[61995]: INFO nova.compute.manager [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Took 20.52 seconds to build instance. [ 981.400923] env[61995]: DEBUG oslo_vmware.api [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-795003, 'name': PowerOffVM_Task, 'duration_secs': 0.269918} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.401678] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 981.401678] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 981.403229] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-075c4420-83da-4917-9350-aef8ea5bb1a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.506629] env[61995]: DEBUG nova.network.neutron [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Successfully updated port: 0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.567497] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52365df3-5f67-dc38-9344-b175ca5422fe, 'name': SearchDatastore_Task, 'duration_secs': 0.012773} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.567814] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.568275] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 981.568755] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.568984] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.569281] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 981.569637] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d67378c-63f8-48e5-bb2d-d1b3a22a1c3a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.582162] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 981.582503] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 981.583925] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c4af245-4434-4aae-80f9-48b0ebdea9cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.588356] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 981.588562] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 981.588729] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Deleting the datastore file [datastore2] 847df13f-8451-49a1-9c6a-7c7c26952940 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.589062] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d446a475-aeeb-482a-aafe-ad368bdaa82c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.594449] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 981.594449] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522f6a12-3d28-3c71-8dd0-9930f29aac68" [ 981.594449] env[61995]: _type = "Task" [ 981.594449] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.602076] env[61995]: DEBUG oslo_vmware.api [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for the task: (returnval){ [ 981.602076] env[61995]: value = "task-795007" [ 981.602076] env[61995]: _type = "Task" [ 981.602076] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.605975] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522f6a12-3d28-3c71-8dd0-9930f29aac68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.615684] env[61995]: DEBUG oslo_vmware.api [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-795007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.683495] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795004, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.834697] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 981.835358] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 981.835751] env[61995]: DEBUG oslo_concurrency.lockutils [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.888387] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4f958dbe-c750-4a1c-8a33-c40612d5ae63 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.034s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.016278] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-9627f278-73d6-45d3-b60d-776e373eef73" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.016609] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-9627f278-73d6-45d3-b60d-776e373eef73" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.016862] env[61995]: DEBUG nova.network.neutron [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 982.055970] env[61995]: DEBUG nova.network.neutron [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Updated VIF entry in instance network info cache for port fd56d4b0-9fa3-445a-a99c-67842d5494e7. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.056085] env[61995]: DEBUG nova.network.neutron [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Updating instance_info_cache with network_info: [{"id": "fd56d4b0-9fa3-445a-a99c-67842d5494e7", "address": "fa:16:3e:d9:01:4b", "network": {"id": "0ac6cb3b-daaf-45e1-b6ce-661de419b804", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1492817454-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9301d14211a1464eb740d9bb745b631a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd56d4b0-9f", "ovs_interfaceid": "fd56d4b0-9fa3-445a-a99c-67842d5494e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.101630] env[61995]: DEBUG nova.network.neutron [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.106910] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522f6a12-3d28-3c71-8dd0-9930f29aac68, 'name': SearchDatastore_Task, 'duration_secs': 0.018439} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.111738] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf11a01e-0422-48d1-8bf1-82faaffe36d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.117561] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 982.117561] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b2648f-723b-2f86-44ad-b9273b3ad352" [ 982.117561] env[61995]: _type = "Task" [ 982.117561] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.120588] env[61995]: DEBUG oslo_vmware.api [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Task: {'id': task-795007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.466531} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.123828] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.124041] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 982.124245] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 982.124421] env[61995]: INFO nova.compute.manager [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Took 1.28 seconds to destroy the instance on the hypervisor. [ 982.124659] env[61995]: DEBUG oslo.service.loopingcall [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.125109] env[61995]: DEBUG nova.compute.manager [-] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.125210] env[61995]: DEBUG nova.network.neutron [-] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 982.132362] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b2648f-723b-2f86-44ad-b9273b3ad352, 'name': SearchDatastore_Task, 'duration_secs': 0.012629} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.132605] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.132868] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 982.133156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.133358] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.133830] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ae9e7bc-4869-4b81-9fee-3b1101c04d63 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.136300] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23d300cc-2c3a-4542-9c19-8983a520328c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.143551] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 982.143551] env[61995]: value = "task-795008" [ 982.143551] env[61995]: _type = "Task" [ 982.143551] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.148384] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.148633] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.149979] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d16259c4-f35e-4b8e-812c-b7e146568789 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.156876] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.160421] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 982.160421] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d03cdb-3817-5811-5dcd-3542d62a908b" [ 982.160421] env[61995]: _type = "Task" [ 982.160421] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.171141] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d03cdb-3817-5811-5dcd-3542d62a908b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.180470] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795004, 'name': CreateVM_Task, 'duration_secs': 0.538271} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.180649] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 982.181336] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.181515] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.181845] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.182137] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7e75ad6-5a54-4103-86ab-047609db304c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.187572] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 982.187572] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52eaab67-d6ac-923c-9a1e-40cabf3efd9b" [ 982.187572] env[61995]: _type = "Task" [ 982.187572] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.196405] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52eaab67-d6ac-923c-9a1e-40cabf3efd9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.559332] env[61995]: DEBUG oslo_concurrency.lockutils [req-3388f826-50d4-41e2-9e42-1df2a42bf6ee req-db4d6264-b083-4ce2-b21a-728b1f294b10 service nova] Releasing lock "refresh_cache-3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.562708] env[61995]: DEBUG nova.network.neutron [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 982.608474] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.658254] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795008, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.674346] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d03cdb-3817-5811-5dcd-3542d62a908b, 'name': SearchDatastore_Task, 'duration_secs': 0.0097} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.675397] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f42ace1-35e6-4ecd-b9c2-a4314cb67303 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.683714] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 982.683714] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52fa1339-a8a9-05b1-38d9-9c80eba0307b" [ 982.683714] env[61995]: _type = "Task" [ 982.683714] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.697074] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fa1339-a8a9-05b1-38d9-9c80eba0307b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.706949] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52eaab67-d6ac-923c-9a1e-40cabf3efd9b, 'name': SearchDatastore_Task, 'duration_secs': 0.014998} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.707382] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.707640] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.708150] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.708295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.708371] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.708652] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f698025-b75e-4b05-bdbc-5ccd09ed1dd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.730762] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.731103] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.731958] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5abf1cb-67d0-4096-8431-db627e3315f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.739380] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 982.739380] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52980e91-77d9-16c3-92cc-1755fb4cd89d" [ 982.739380] env[61995]: _type = "Task" [ 982.739380] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.744460] env[61995]: DEBUG nova.network.neutron [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Updating instance_info_cache with network_info: [{"id": "0bd3d577-8a66-4cce-bc58-1d82dd64bbf2", "address": "fa:16:3e:28:c1:66", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd3d577-8a", "ovs_interfaceid": "0bd3d577-8a66-4cce-bc58-1d82dd64bbf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.749458] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52980e91-77d9-16c3-92cc-1755fb4cd89d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.860798] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Received event network-changed-592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.861068] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Refreshing instance network info cache due to event network-changed-592fb7d9-e1e2-4233-80f7-6a263f59fac2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 982.861287] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Acquiring lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.861471] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Acquired lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.861682] env[61995]: DEBUG nova.network.neutron [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Refreshing network info cache for port 592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 982.910398] env[61995]: DEBUG nova.network.neutron [-] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.101402] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.101603] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.156109] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.710062} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.156432] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 983.156658] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.157626] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20f550ee-2670-4492-be5a-3dce61d46f55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.166359] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 983.166359] env[61995]: value = "task-795009" [ 983.166359] env[61995]: _type = "Task" [ 983.166359] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.175749] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795009, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.195035] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fa1339-a8a9-05b1-38d9-9c80eba0307b, 'name': SearchDatastore_Task, 'duration_secs': 0.050727} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.195399] env[61995]: DEBUG oslo_concurrency.lockutils [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.195765] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk. {{(pid=61995) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 983.196507] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5f4af4e-c2ea-44d9-a5da-867955ebce0d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.204291] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 983.204291] env[61995]: value = "task-795010" [ 983.204291] env[61995]: _type = "Task" [ 983.204291] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.216037] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795010, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.252262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-9627f278-73d6-45d3-b60d-776e373eef73" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.252743] env[61995]: DEBUG nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Instance network_info: |[{"id": "0bd3d577-8a66-4cce-bc58-1d82dd64bbf2", "address": "fa:16:3e:28:c1:66", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd3d577-8a", "ovs_interfaceid": "0bd3d577-8a66-4cce-bc58-1d82dd64bbf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 983.253151] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52980e91-77d9-16c3-92cc-1755fb4cd89d, 'name': SearchDatastore_Task, 'duration_secs': 0.024585} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.253592] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:c1:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ada35c98-01a9-4352-98e4-1d20ba31f928', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0bd3d577-8a66-4cce-bc58-1d82dd64bbf2', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.264301] env[61995]: DEBUG oslo.service.loopingcall [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.265372] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 983.265657] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa001f5f-61e6-469f-b881-f9e9ef038484 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.268874] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55d249e6-6781-44fb-9107-cdbecb6770ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.290772] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 983.290772] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ad35fa-d7e6-8b2e-1cd4-be72740e85d9" [ 983.290772] env[61995]: _type = "Task" [ 983.290772] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.297118] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 983.297118] env[61995]: value = "task-795011" [ 983.297118] env[61995]: _type = "Task" [ 983.297118] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.305841] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ad35fa-d7e6-8b2e-1cd4-be72740e85d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.312995] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795011, 'name': CreateVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.413560] env[61995]: INFO nova.compute.manager [-] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Took 1.29 seconds to deallocate network for instance. [ 983.604141] env[61995]: DEBUG nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.667260] env[61995]: DEBUG nova.network.neutron [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Updated VIF entry in instance network info cache for port 592fb7d9-e1e2-4233-80f7-6a263f59fac2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 983.667672] env[61995]: DEBUG nova.network.neutron [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Updating instance_info_cache with network_info: [{"id": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "address": "fa:16:3e:79:c1:cd", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap592fb7d9-e1", "ovs_interfaceid": "592fb7d9-e1e2-4233-80f7-6a263f59fac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.681912] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795009, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078792} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.683190] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 983.684536] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b204f489-c122-40a8-a15c-213820d41ab2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.713450] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.714217] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b67d8655-56cd-4087-9bc1-5bfd46d22e1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.740018] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795010, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.741400] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 983.741400] env[61995]: value = "task-795012" [ 983.741400] env[61995]: _type = "Task" [ 983.741400] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.800796] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ad35fa-d7e6-8b2e-1cd4-be72740e85d9, 'name': SearchDatastore_Task, 'duration_secs': 0.018361} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.804228] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.804228] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4/3a2adf56-8071-476d-8cf8-5f6ece6a1ac4.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.804389] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0df76cfa-3a2d-4c54-9841-f2456df65fe1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.813429] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795011, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.814820] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 983.814820] env[61995]: value = "task-795013" [ 983.814820] env[61995]: _type = "Task" [ 983.814820] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.823643] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.920179] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.920485] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.920822] env[61995]: DEBUG nova.objects.instance [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lazy-loading 'resources' on Instance uuid 847df13f-8451-49a1-9c6a-7c7c26952940 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.126452] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d6a44f-02ff-4d94-add0-7135799e96d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.130557] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.146855] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance '229d966c-b70d-4a51-a176-2e88488d5c4b' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 984.175971] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Releasing lock "refresh_cache-ba89ee63-e293-47e1-90ab-7b8e72dd1b50" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.175971] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Received event network-vif-plugged-0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.175971] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Acquiring lock "9627f278-73d6-45d3-b60d-776e373eef73-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.175971] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Lock "9627f278-73d6-45d3-b60d-776e373eef73-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.176275] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Lock "9627f278-73d6-45d3-b60d-776e373eef73-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.176315] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] No waiting events found dispatching network-vif-plugged-0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.177022] env[61995]: WARNING nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Received unexpected event network-vif-plugged-0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 for instance with vm_state building and task_state spawning. [ 984.177022] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Received event network-changed-0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.177022] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Refreshing instance network info cache due to event network-changed-0bd3d577-8a66-4cce-bc58-1d82dd64bbf2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 984.177022] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Acquiring lock "refresh_cache-9627f278-73d6-45d3-b60d-776e373eef73" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.177275] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Acquired lock "refresh_cache-9627f278-73d6-45d3-b60d-776e373eef73" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.177275] env[61995]: DEBUG nova.network.neutron [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Refreshing network info cache for port 0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.225156] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795010, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579583} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.225424] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk. [ 984.226217] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e27857-bad5-4b57-acb7-86ca69a2049e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.251367] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.254981] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a0d3d4e-ecfc-4e69-98b4-ae0a503af0bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.275737] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795012, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.277034] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 984.277034] env[61995]: value = "task-795014" [ 984.277034] env[61995]: _type = "Task" [ 984.277034] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.284804] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795014, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.309580] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795011, 'name': CreateVM_Task, 'duration_secs': 0.950338} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.309752] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 984.310437] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.310621] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.310942] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.311200] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb5226b9-4df7-4b6b-8c6c-0e5a3d83a7a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.315857] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 984.315857] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5207c01d-d130-2861-8815-aec36869fdd2" [ 984.315857] env[61995]: _type = "Task" [ 984.315857] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.326844] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.329685] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5207c01d-d130-2861-8815-aec36869fdd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.648181] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94ca763-41cb-443d-9fdc-cc3b28a62a83 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.652502] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.652808] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-812feb5d-61cc-4c89-8684-f47935eff7d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.665072] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a85c67-2e69-461e-9d17-8de47cc61341 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.668051] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 984.668051] env[61995]: value = "task-795015" [ 984.668051] env[61995]: _type = "Task" [ 984.668051] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.703116] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1505fd57-4dae-4b40-8a40-dcef8f7b4a43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.711420] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 984.711669] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance '229d966c-b70d-4a51-a176-2e88488d5c4b' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 984.721532] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2c0802-f681-466f-a65a-0fbfe915ac24 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.738367] env[61995]: DEBUG nova.compute.provider_tree [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.758757] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795012, 'name': ReconfigVM_Task, 'duration_secs': 1.014606} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.759110] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfigured VM instance instance-0000004c to attach disk [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31/bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.760571] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_format': None, 'device_type': 'disk', 'encryption_secret_uuid': None, 'device_name': '/dev/sda', 'encryption_options': None, 'guest_format': None, 'encrypted': False, 'boot_index': 0, 'size': 0, 'disk_bus': None, 'image_id': 'c1739132-aff3-4644-9522-fc02cf765bc8'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'}, 'attachment_id': '9d764864-b88f-479c-85d5-b378581dd175', 'guest_format': None, 'boot_index': None, 'disk_bus': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=61995) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 984.760808] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 984.761037] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 984.762192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9562d758-fabf-46d4-8d04-c1f6c4b3f912 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.785269] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cff334-3adc-4e75-86e0-21bce06a7ab1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.810049] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795014, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.818138] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068/volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.821093] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06aab628-4500-4100-aea6-fdd2e95ebf05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.849268] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795013, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.854323] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 984.854323] env[61995]: value = "task-795016" [ 984.854323] env[61995]: _type = "Task" [ 984.854323] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.854755] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5207c01d-d130-2861-8815-aec36869fdd2, 'name': SearchDatastore_Task, 'duration_secs': 0.039129} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.855061] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.855385] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 984.855682] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.856485] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.856485] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 984.859466] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2d6368d-e71c-4625-9d46-8bed1e73ee27 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.868204] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795016, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.870482] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.870676] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 984.871575] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb0f5d54-29da-40c6-9629-5e1842a4db23 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.877621] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 984.877621] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52392628-6e49-e409-5e17-29d2bef76b0b" [ 984.877621] env[61995]: _type = "Task" [ 984.877621] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.886367] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52392628-6e49-e409-5e17-29d2bef76b0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.031338] env[61995]: DEBUG nova.network.neutron [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Updated VIF entry in instance network info cache for port 0bd3d577-8a66-4cce-bc58-1d82dd64bbf2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.031709] env[61995]: DEBUG nova.network.neutron [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Updating instance_info_cache with network_info: [{"id": "0bd3d577-8a66-4cce-bc58-1d82dd64bbf2", "address": "fa:16:3e:28:c1:66", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd3d577-8a", "ovs_interfaceid": "0bd3d577-8a66-4cce-bc58-1d82dd64bbf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.222352] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.222601] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.222767] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.222956] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.223123] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.223289] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.223492] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.223666] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.223845] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.224020] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.224290] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.229367] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c20235c1-8fa0-4991-96f6-67694d7552c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.241602] env[61995]: DEBUG nova.scheduler.client.report [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.246074] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 985.246074] env[61995]: value = "task-795017" [ 985.246074] env[61995]: _type = "Task" [ 985.246074] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.254956] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795017, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.291939] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795014, 'name': ReconfigVM_Task, 'duration_secs': 0.786695} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.292258] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.293094] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b2c229-abe7-4d4a-906f-7d7c13ec8a05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.318434] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-370a066a-e81e-4bd8-8cd4-99430bb7c988 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.339325] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795013, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.047685} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.340522] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4/3a2adf56-8071-476d-8cf8-5f6ece6a1ac4.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 985.340760] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.341081] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 985.341081] env[61995]: value = "task-795018" [ 985.341081] env[61995]: _type = "Task" [ 985.341081] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.341279] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-387852e0-60b1-4f4d-bdd2-fc58d13013ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.351687] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795018, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.352878] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 985.352878] env[61995]: value = "task-795019" [ 985.352878] env[61995]: _type = "Task" [ 985.352878] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.365424] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.368341] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795019, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.387369] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52392628-6e49-e409-5e17-29d2bef76b0b, 'name': SearchDatastore_Task, 'duration_secs': 0.014976} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.388284] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ed12499-881d-48de-80b2-c50576bfa4d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.394990] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 985.394990] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5227a9f3-94ce-ad9c-22e8-55a21f7484ac" [ 985.394990] env[61995]: _type = "Task" [ 985.394990] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.403591] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5227a9f3-94ce-ad9c-22e8-55a21f7484ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.534287] env[61995]: DEBUG oslo_concurrency.lockutils [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] Releasing lock "refresh_cache-9627f278-73d6-45d3-b60d-776e373eef73" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.534589] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Received event network-vif-deleted-4554d960-3f34-464e-abd8-8f16166194b3 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.534793] env[61995]: INFO nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Neutron deleted interface 4554d960-3f34-464e-abd8-8f16166194b3; detaching it from the instance and deleting it from the info cache [ 985.534971] env[61995]: DEBUG nova.network.neutron [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.748371] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.751979] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.621s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.754356] env[61995]: INFO nova.compute.claims [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.771789] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795017, 'name': ReconfigVM_Task, 'duration_secs': 0.406522} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.773600] env[61995]: INFO nova.scheduler.client.report [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Deleted allocations for instance 847df13f-8451-49a1-9c6a-7c7c26952940 [ 985.775083] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance '229d966c-b70d-4a51-a176-2e88488d5c4b' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 985.856560] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795018, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.869568] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.872441] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795019, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070786} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.872704] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.873482] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4265c3-1ff9-4ab1-ad4f-aa7cd8f0edeb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.896376] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4/3a2adf56-8071-476d-8cf8-5f6ece6a1ac4.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.896714] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-357adb56-9d18-4cfd-9781-1c29850eab85 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.920969] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5227a9f3-94ce-ad9c-22e8-55a21f7484ac, 'name': SearchDatastore_Task, 'duration_secs': 0.0132} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.922245] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.922608] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 985.922933] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 985.922933] env[61995]: value = "task-795020" [ 985.922933] env[61995]: _type = "Task" [ 985.922933] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.923144] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c55e358-c52c-45be-866a-a911d26e53be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.935870] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795020, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.937246] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 985.937246] env[61995]: value = "task-795021" [ 985.937246] env[61995]: _type = "Task" [ 985.937246] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.945732] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795021, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.037679] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-13d2d2fd-fe67-40c0-bc7f-d784e7570c7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.047794] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef2fce8-bfdb-48d5-90ad-e68f491c809e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.081457] env[61995]: DEBUG nova.compute.manager [req-05bcf3a8-5c5f-469c-9be8-99e7a096a3d6 req-0c042fe6-da9a-47b0-a589-880b118b0832 service nova] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Detach interface failed, port_id=4554d960-3f34-464e-abd8-8f16166194b3, reason: Instance 847df13f-8451-49a1-9c6a-7c7c26952940 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 986.284326] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.284600] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.284806] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.285015] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.285181] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.285329] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.285537] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.285700] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.285871] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.286106] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.286295] env[61995]: DEBUG nova.virt.hardware [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.291923] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Reconfiguring VM instance instance-00000040 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 986.294537] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a8e5e94-b2f1-4f63-8da3-a7eec301a16d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.309387] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cfa4254d-0266-484b-8869-9bd2857e1a1f tempest-FloatingIPsAssociationNegativeTestJSON-1332852081 tempest-FloatingIPsAssociationNegativeTestJSON-1332852081-project-member] Lock "847df13f-8451-49a1-9c6a-7c7c26952940" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.470s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.317929] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 986.317929] env[61995]: value = "task-795022" [ 986.317929] env[61995]: _type = "Task" [ 986.317929] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.328173] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.353906] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795018, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.366606] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795016, 'name': ReconfigVM_Task, 'duration_secs': 1.166849} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.366999] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfigured VM instance instance-0000004c to attach disk [datastore2] volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068/volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.371951] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6eee04ec-60a4-465f-b7a7-e8ce58fa2606 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.389255] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 986.389255] env[61995]: value = "task-795023" [ 986.389255] env[61995]: _type = "Task" [ 986.389255] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.400351] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795023, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.436471] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795020, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.448380] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795021, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.829157] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795022, 'name': ReconfigVM_Task, 'duration_secs': 0.251159} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.831779] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Reconfigured VM instance instance-00000040 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 986.832451] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9362be-9499-4f6a-877f-1742e2a3fa62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.857235] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 229d966c-b70d-4a51-a176-2e88488d5c4b/229d966c-b70d-4a51-a176-2e88488d5c4b.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.863223] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abbb54ef-6435-4ca1-b23c-2f3b4af26fe4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.883266] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795018, 'name': ReconfigVM_Task, 'duration_secs': 1.015732} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.884593] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.884827] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 986.884827] env[61995]: value = "task-795024" [ 986.884827] env[61995]: _type = "Task" [ 986.884827] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.887299] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a53930e5-fc5b-4225-ba86-183f6380f590 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.897151] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 986.897151] env[61995]: value = "task-795025" [ 986.897151] env[61995]: _type = "Task" [ 986.897151] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.906459] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795024, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.911711] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795023, 'name': ReconfigVM_Task, 'duration_secs': 0.218842} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.912188] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 986.913397] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a145cbd-005c-4e85-b4c2-c26cbbd2dec8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.918086] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.924856] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 986.924856] env[61995]: value = "task-795026" [ 986.924856] env[61995]: _type = "Task" [ 986.924856] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.939840] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795020, 'name': ReconfigVM_Task, 'duration_secs': 0.544977} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.948383] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4/3a2adf56-8071-476d-8cf8-5f6ece6a1ac4.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.949141] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795026, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.949595] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb82b1c9-a1ec-4a27-89ad-ba868c19e60f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.957128] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795021, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.824639} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.958459] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 986.959046] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.959046] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 986.959046] env[61995]: value = "task-795027" [ 986.959046] env[61995]: _type = "Task" [ 986.959046] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.959378] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-209cc61e-6006-41c4-b519-09a669d9dfaf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.974134] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795027, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.975554] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 986.975554] env[61995]: value = "task-795028" [ 986.975554] env[61995]: _type = "Task" [ 986.975554] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.987337] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795028, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.034541] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7984495c-5118-487f-8b66-8f546ca37a7f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.043250] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5224828-8f67-4c12-be3d-bbc4e658b3c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.078461] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74c3a7f-f2a2-46e9-8061-8c8375695a46 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.087309] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9b4e98-2b7f-4338-8546-29407311fef8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.104434] env[61995]: DEBUG nova.compute.provider_tree [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.399318] env[61995]: DEBUG oslo_vmware.api [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795024, 'name': ReconfigVM_Task, 'duration_secs': 0.32474} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.405097] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 229d966c-b70d-4a51-a176-2e88488d5c4b/229d966c-b70d-4a51-a176-2e88488d5c4b.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.405097] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance '229d966c-b70d-4a51-a176-2e88488d5c4b' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.411866] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795025, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.439766] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795026, 'name': Rename_Task, 'duration_secs': 0.180198} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.440234] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.441626] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6eb405f8-9037-4607-adfb-3d76861a71c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.448794] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 987.448794] env[61995]: value = "task-795029" [ 987.448794] env[61995]: _type = "Task" [ 987.448794] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.458105] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.476554] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795027, 'name': Rename_Task, 'duration_secs': 0.167887} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.480086] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.480398] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85d24a32-1909-4d04-8c06-f0f1892230c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.489178] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795028, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084183} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.490577] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.490931] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 987.490931] env[61995]: value = "task-795030" [ 987.490931] env[61995]: _type = "Task" [ 987.490931] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.491648] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ca1913-ab84-4fd4-a1e3-045468ed6f51 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.502333] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795030, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.520168] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.520815] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34bb328f-bed6-4a2b-af48-f7ea9d7ca7ca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.542947] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 987.542947] env[61995]: value = "task-795031" [ 987.542947] env[61995]: _type = "Task" [ 987.542947] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.552551] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795031, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.607435] env[61995]: DEBUG nova.scheduler.client.report [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.911386] env[61995]: DEBUG oslo_vmware.api [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795025, 'name': PowerOnVM_Task, 'duration_secs': 0.874326} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.911386] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3ca925-1814-41e0-b638-2c4387083c29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.913901] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.916624] env[61995]: DEBUG nova.compute.manager [None req-335d1ca2-2a3f-4610-9193-693a5fd0ce0b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.917354] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52705843-2892-44ca-b449-d9ddc1f2e02c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.936790] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef92013-09c8-4143-9a49-9502ddead604 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.962577] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance '229d966c-b70d-4a51-a176-2e88488d5c4b' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.977107] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795029, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.006256] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795030, 'name': PowerOnVM_Task} progress is 76%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.053494] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795031, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.112950] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.113579] env[61995]: DEBUG nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.478563] env[61995]: DEBUG oslo_vmware.api [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795029, 'name': PowerOnVM_Task, 'duration_secs': 0.986144} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.479904] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.479904] env[61995]: DEBUG nova.compute.manager [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.480806] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078d1aa2-0178-4f68-8f80-c8bd7644190a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.506173] env[61995]: DEBUG oslo_vmware.api [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795030, 'name': PowerOnVM_Task, 'duration_secs': 0.960733} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.506173] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.506398] env[61995]: INFO nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Took 10.63 seconds to spawn the instance on the hypervisor. [ 988.506398] env[61995]: DEBUG nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.507485] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73abb2d-b539-4705-af36-0b7ab2dcece2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.548243] env[61995]: DEBUG nova.network.neutron [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Port cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 988.568538] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795031, 'name': ReconfigVM_Task, 'duration_secs': 0.640903} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.568821] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.569802] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8500887-727f-4649-a529-1a5313940d51 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.577068] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 988.577068] env[61995]: value = "task-795032" [ 988.577068] env[61995]: _type = "Task" [ 988.577068] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.586067] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795032, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.619411] env[61995]: DEBUG nova.compute.utils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.620955] env[61995]: DEBUG nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 988.621179] env[61995]: DEBUG nova.network.neutron [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 988.659654] env[61995]: DEBUG nova.policy [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caf9bbbd848d4d7ba53ba3bd553943e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2018aae65cf4bf98b5ac82519201cf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 988.900031] env[61995]: INFO nova.compute.manager [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Unrescuing [ 988.900031] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.900031] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.900031] env[61995]: DEBUG nova.network.neutron [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 988.925172] env[61995]: DEBUG nova.network.neutron [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Successfully created port: 70414355-248a-4beb-b64e-aedd609cd31b {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.009022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.009022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.009022] env[61995]: DEBUG nova.objects.instance [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 989.030376] env[61995]: INFO nova.compute.manager [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Took 21.33 seconds to build instance. [ 989.088315] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795032, 'name': Rename_Task, 'duration_secs': 0.291809} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.088628] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 989.088871] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-281710cb-0f46-477a-b0f7-f37d10db475f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.096799] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 989.096799] env[61995]: value = "task-795033" [ 989.096799] env[61995]: _type = "Task" [ 989.096799] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.106224] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.125598] env[61995]: DEBUG nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.533231] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7d361081-2be7-4938-85bf-c73c4f3da51c tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.839s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.582445] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.582700] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.582880] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.608937] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795033, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.761275] env[61995]: DEBUG nova.network.neutron [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updating instance_info_cache with network_info: [{"id": "2327f433-5c63-4d83-8050-6661d9ad60eb", "address": "fa:16:3e:d0:20:ad", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2327f433-5c", "ovs_interfaceid": "2327f433-5c63-4d83-8050-6661d9ad60eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.025071] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3bd9a827-55f5-4973-b8f6-9c2b376da3d9 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.111028] env[61995]: DEBUG oslo_vmware.api [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795033, 'name': PowerOnVM_Task, 'duration_secs': 0.898535} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.111369] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.111616] env[61995]: INFO nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Took 9.83 seconds to spawn the instance on the hypervisor. [ 990.111837] env[61995]: DEBUG nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.112703] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeb3c39-e04a-4b3e-b9c2-22ed4ae2e8a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.139130] env[61995]: DEBUG nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.170426] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.170680] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.170842] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.171057] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.171211] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.171373] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.171614] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.171782] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.171959] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.172159] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.173207] env[61995]: DEBUG nova.virt.hardware [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.173277] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7e6145-c3bd-4659-aa66-ec5c32ac9bbe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.185019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6cdc37-2521-4136-b0f2-267143fdd804 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.267180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.267853] env[61995]: DEBUG nova.objects.instance [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lazy-loading 'flavor' on Instance uuid 24b974c0-9567-46a3-96f6-fac1e602f46e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.458774] env[61995]: DEBUG oslo_concurrency.lockutils [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.458774] env[61995]: DEBUG oslo_concurrency.lockutils [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.499352] env[61995]: DEBUG nova.compute.manager [req-133b2a13-5675-4e0a-ac8f-23bd297beb19 req-5404426d-f063-4e88-a027-b3fa7c1e3868 service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Received event network-vif-plugged-70414355-248a-4beb-b64e-aedd609cd31b {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.499352] env[61995]: DEBUG oslo_concurrency.lockutils [req-133b2a13-5675-4e0a-ac8f-23bd297beb19 req-5404426d-f063-4e88-a027-b3fa7c1e3868 service nova] Acquiring lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.499352] env[61995]: DEBUG oslo_concurrency.lockutils [req-133b2a13-5675-4e0a-ac8f-23bd297beb19 req-5404426d-f063-4e88-a027-b3fa7c1e3868 service nova] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.499352] env[61995]: DEBUG oslo_concurrency.lockutils [req-133b2a13-5675-4e0a-ac8f-23bd297beb19 req-5404426d-f063-4e88-a027-b3fa7c1e3868 service nova] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.499352] env[61995]: DEBUG nova.compute.manager [req-133b2a13-5675-4e0a-ac8f-23bd297beb19 req-5404426d-f063-4e88-a027-b3fa7c1e3868 service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] No waiting events found dispatching network-vif-plugged-70414355-248a-4beb-b64e-aedd609cd31b {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 990.499352] env[61995]: WARNING nova.compute.manager [req-133b2a13-5675-4e0a-ac8f-23bd297beb19 req-5404426d-f063-4e88-a027-b3fa7c1e3868 service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Received unexpected event network-vif-plugged-70414355-248a-4beb-b64e-aedd609cd31b for instance with vm_state building and task_state spawning. [ 990.633655] env[61995]: INFO nova.compute.manager [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Took 19.16 seconds to build instance. [ 990.642585] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.642585] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.642585] env[61995]: DEBUG nova.network.neutron [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 990.659506] env[61995]: DEBUG nova.network.neutron [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Successfully updated port: 70414355-248a-4beb-b64e-aedd609cd31b {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.779068] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aaaf228-e8b4-4e85-b5a9-104c400c8421 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.807239] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.807239] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c00cfa6-7d19-439d-8892-4262fa5127a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.815116] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 990.815116] env[61995]: value = "task-795034" [ 990.815116] env[61995]: _type = "Task" [ 990.815116] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.826934] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795034, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.961318] env[61995]: INFO nova.compute.manager [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Detaching volume caeb58b4-c8a5-46b3-9f21-57540bd5b068 [ 991.020505] env[61995]: INFO nova.virt.block_device [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Attempting to driver detach volume caeb58b4-c8a5-46b3-9f21-57540bd5b068 from mountpoint /dev/sdb [ 991.020780] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 991.020995] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 991.021971] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd86a1de-6e93-4add-acb6-1582555a0444 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.047050] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b60791-9f8c-49a7-b19a-adaf9bfc6ce8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.056515] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46cbb25-086f-4f18-b405-356ea63e5993 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.082217] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a227294-e7cc-400d-bdcc-7e5e64842743 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.096577] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] The volume has not been displaced from its original location: [datastore2] volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068/volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 991.102213] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 991.102990] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48de2c0f-041e-4f46-8f0a-63e4ecba8f4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.123904] env[61995]: DEBUG oslo_vmware.api [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 991.123904] env[61995]: value = "task-795035" [ 991.123904] env[61995]: _type = "Task" [ 991.123904] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.135559] env[61995]: DEBUG oslo_concurrency.lockutils [None req-236774ba-5457-49b3-bc1a-e0b813715860 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "9627f278-73d6-45d3-b60d-776e373eef73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.666s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.140424] env[61995]: DEBUG oslo_vmware.api [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795035, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.164985] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "refresh_cache-349667a3-6311-4c9d-bad2-ef4a5d1c52c8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.165229] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "refresh_cache-349667a3-6311-4c9d-bad2-ef4a5d1c52c8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.165361] env[61995]: DEBUG nova.network.neutron [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.333848] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795034, 'name': PowerOffVM_Task, 'duration_secs': 0.485985} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.334358] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.334717] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.334911] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.335192] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.335370] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.338827] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.346099] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Reconfiguring VM instance instance-0000004f to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 991.347126] env[61995]: INFO nova.compute.manager [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Terminating instance [ 991.348451] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9fc03db-115b-4ece-904e-3bed668228d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.365353] env[61995]: DEBUG nova.compute.manager [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.370022] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 991.370022] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812ea12e-5499-4fc2-aa4a-35e6f1dacf43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.378671] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 991.380288] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa1c5013-28e4-46b0-9adb-730abf66190b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.382035] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 991.382035] env[61995]: value = "task-795036" [ 991.382035] env[61995]: _type = "Task" [ 991.382035] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.389752] env[61995]: DEBUG oslo_vmware.api [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 991.389752] env[61995]: value = "task-795037" [ 991.389752] env[61995]: _type = "Task" [ 991.389752] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.398233] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795036, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.404670] env[61995]: DEBUG oslo_vmware.api [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.638164] env[61995]: DEBUG oslo_vmware.api [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795035, 'name': ReconfigVM_Task, 'duration_secs': 0.378122} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.638164] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 991.642153] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4178e293-d87c-4c5f-bf0f-9e70a9fec376 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.658863] env[61995]: DEBUG nova.network.neutron [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.661971] env[61995]: DEBUG oslo_vmware.api [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 991.661971] env[61995]: value = "task-795038" [ 991.661971] env[61995]: _type = "Task" [ 991.661971] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.674383] env[61995]: DEBUG oslo_vmware.api [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795038, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.719616] env[61995]: DEBUG nova.network.neutron [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 991.893771] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795036, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.903893] env[61995]: DEBUG oslo_vmware.api [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795037, 'name': PowerOffVM_Task, 'duration_secs': 0.50613} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.903893] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.904196] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.904510] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba7d4f4d-c3cc-4ed2-9fde-54e8fcbadd5a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.017811] env[61995]: DEBUG nova.network.neutron [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Updating instance_info_cache with network_info: [{"id": "70414355-248a-4beb-b64e-aedd609cd31b", "address": "fa:16:3e:e6:c3:41", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70414355-24", "ovs_interfaceid": "70414355-248a-4beb-b64e-aedd609cd31b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.166156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.181278] env[61995]: DEBUG oslo_vmware.api [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795038, 'name': ReconfigVM_Task, 'duration_secs': 0.240381} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.181595] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185376', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'name': 'volume-caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31', 'attached_at': '', 'detached_at': '', 'volume_id': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068', 'serial': 'caeb58b4-c8a5-46b3-9f21-57540bd5b068'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 992.194254] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 992.194254] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 992.194367] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleting the datastore file [datastore2] 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.194663] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a185a3e-1657-4269-943f-926142609e7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.203109] env[61995]: DEBUG oslo_vmware.api [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for the task: (returnval){ [ 992.203109] env[61995]: value = "task-795040" [ 992.203109] env[61995]: _type = "Task" [ 992.203109] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.215642] env[61995]: DEBUG oslo_vmware.api [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.253252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0bcc3a97-3637-48b9-9871-e56279cdd4ee tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.253252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0bcc3a97-3637-48b9-9871-e56279cdd4ee tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.253455] env[61995]: DEBUG nova.objects.instance [None req-0bcc3a97-3637-48b9-9871-e56279cdd4ee tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'flavor' on Instance uuid 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.300017] env[61995]: INFO nova.compute.manager [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Rebuilding instance [ 992.352042] env[61995]: DEBUG nova.compute.manager [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.352908] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2081cd33-316e-4fc8-b64d-05a66ea9ba0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.393407] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795036, 'name': ReconfigVM_Task, 'duration_secs': 0.52149} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.393739] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Reconfigured VM instance instance-0000004f to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 992.393958] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 992.394229] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-530cd74e-d093-425e-b67c-c429f50b9759 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.404577] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 992.404577] env[61995]: value = "task-795041" [ 992.404577] env[61995]: _type = "Task" [ 992.404577] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.414139] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795041, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.522221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "refresh_cache-349667a3-6311-4c9d-bad2-ef4a5d1c52c8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.522590] env[61995]: DEBUG nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Instance network_info: |[{"id": "70414355-248a-4beb-b64e-aedd609cd31b", "address": "fa:16:3e:e6:c3:41", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70414355-24", "ovs_interfaceid": "70414355-248a-4beb-b64e-aedd609cd31b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.523048] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:c3:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70414355-248a-4beb-b64e-aedd609cd31b', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.532799] env[61995]: DEBUG oslo.service.loopingcall [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.533103] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 992.533331] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3efed37e-404c-4c11-8571-193ab9b7d7c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.553199] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.553199] env[61995]: value = "task-795042" [ 992.553199] env[61995]: _type = "Task" [ 992.553199] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.562698] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795042, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.700989] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed980c2d-2893-4289-bcc1-44a2c057cd03 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.715728] env[61995]: DEBUG oslo_vmware.api [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Task: {'id': task-795040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.33669} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.728694] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.729018] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.729134] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.729370] env[61995]: INFO nova.compute.manager [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Took 1.36 seconds to destroy the instance on the hypervisor. [ 992.729641] env[61995]: DEBUG oslo.service.loopingcall [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.730068] env[61995]: DEBUG nova.compute.manager [-] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.730155] env[61995]: DEBUG nova.network.neutron [-] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 992.732977] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6701aee6-2b74-4ae9-8f2e-e27c0e903292 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.741913] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance '229d966c-b70d-4a51-a176-2e88488d5c4b' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 992.757963] env[61995]: DEBUG nova.objects.instance [None req-0bcc3a97-3637-48b9-9871-e56279cdd4ee tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'pci_requests' on Instance uuid 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.761121] env[61995]: DEBUG nova.objects.instance [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'flavor' on Instance uuid bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.870900] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 992.871296] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fec310b6-1c3b-46a9-927f-50b9f4688f7a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.879915] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 992.879915] env[61995]: value = "task-795043" [ 992.879915] env[61995]: _type = "Task" [ 992.879915] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.889464] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.662232] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fffc6a80-c5cb-4926-a4af-2c3894526aa2 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance '229d966c-b70d-4a51-a176-2e88488d5c4b' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 993.665912] env[61995]: DEBUG nova.objects.base [None req-0bcc3a97-3637-48b9-9871-e56279cdd4ee tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Object Instance<453f46c7-5bad-4ca2-b228-f76e62fbd03e> lazy-loaded attributes: flavor,pci_requests {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 993.666113] env[61995]: DEBUG nova.network.neutron [None req-0bcc3a97-3637-48b9-9871-e56279cdd4ee tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 993.672024] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.672024] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.672024] env[61995]: DEBUG nova.compute.manager [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Received event network-changed-70414355-248a-4beb-b64e-aedd609cd31b {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 993.672024] env[61995]: DEBUG nova.compute.manager [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Refreshing instance network info cache due to event network-changed-70414355-248a-4beb-b64e-aedd609cd31b. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 993.672364] env[61995]: DEBUG oslo_concurrency.lockutils [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] Acquiring lock "refresh_cache-349667a3-6311-4c9d-bad2-ef4a5d1c52c8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.672364] env[61995]: DEBUG oslo_concurrency.lockutils [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] Acquired lock "refresh_cache-349667a3-6311-4c9d-bad2-ef4a5d1c52c8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.672481] env[61995]: DEBUG nova.network.neutron [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Refreshing network info cache for port 70414355-248a-4beb-b64e-aedd609cd31b {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.693556] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795042, 'name': CreateVM_Task, 'duration_secs': 0.361649} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.698355] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 993.698695] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795043, 'name': PowerOffVM_Task, 'duration_secs': 0.362586} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.699283] env[61995]: DEBUG oslo_vmware.api [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795041, 'name': PowerOnVM_Task, 'duration_secs': 0.457427} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.699859] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.700025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.700354] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.700668] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 993.700831] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 993.702348] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 993.702348] env[61995]: DEBUG nova.compute.manager [None req-e755bfe0-5d57-4fbb-9703-a6ee9a7fe3cb tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.702348] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48b771b7-281e-4ea0-8e2e-e08faca0acd2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.704159] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448040a2-073b-45cd-92fd-32f5fec8f60b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.707540] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fdaf252-3209-4136-8736-289181c17223 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.714626] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 993.714626] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522057b9-d51c-7f42-c3bb-6f05f0144f87" [ 993.714626] env[61995]: _type = "Task" [ 993.714626] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.718615] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 993.724986] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d81ecc9-cc41-49fc-83af-846d408264c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.733265] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522057b9-d51c-7f42-c3bb-6f05f0144f87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.812180] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 993.812334] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 993.812443] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleting the datastore file [datastore1] 9627f278-73d6-45d3-b60d-776e373eef73 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.812721] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fd9eba4-0945-4b84-a77f-a026811bf1c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.818544] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0bcc3a97-3637-48b9-9871-e56279cdd4ee tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.565s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.821937] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 993.821937] env[61995]: value = "task-795045" [ 993.821937] env[61995]: _type = "Task" [ 993.821937] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.830609] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795045, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.084901] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.084901] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.176905] env[61995]: DEBUG nova.network.neutron [-] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.181663] env[61995]: INFO nova.compute.manager [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Detaching volume 908f434d-3808-42d4-a891-104337fb5452 [ 994.186647] env[61995]: DEBUG oslo_concurrency.lockutils [None req-db326a05-9d14-4653-bb2a-d02161256632 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.729s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.224676] env[61995]: INFO nova.virt.block_device [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Attempting to driver detach volume 908f434d-3808-42d4-a891-104337fb5452 from mountpoint /dev/sdb [ 994.225016] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 994.225268] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185367', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'name': 'volume-908f434d-3808-42d4-a891-104337fb5452', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e835ac53-29af-4bd0-b186-5c6270ccf760', 'attached_at': '', 'detached_at': '', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'serial': '908f434d-3808-42d4-a891-104337fb5452'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 994.226531] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57869e9d-e95c-42a1-9e96-f93834950535 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.236871] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522057b9-d51c-7f42-c3bb-6f05f0144f87, 'name': SearchDatastore_Task, 'duration_secs': 0.028107} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.238650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.239380] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.239380] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.239466] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.240028] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.259416] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4500e31-48b7-4dc9-8ded-1cf64be00135 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.262630] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbc5f75-d7dc-4ac4-9311-4fdbb2540ff0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.273540] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a519f7-c89c-4ca0-8caa-faf3d4336bbe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.277024] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.277024] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 994.277024] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4445b45d-ae73-48c1-aeaf-ff1a09634fe0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.298144] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64df7626-4555-4a5e-a721-e3b7ec0c44ce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.299710] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 994.299710] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5228b3ea-6555-7a2b-b2b9-4e6dab7767bd" [ 994.299710] env[61995]: _type = "Task" [ 994.299710] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.314267] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] The volume has not been displaced from its original location: [datastore1] volume-908f434d-3808-42d4-a891-104337fb5452/volume-908f434d-3808-42d4-a891-104337fb5452.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 994.319212] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.327470] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f88fcd50-ed32-4576-8a5a-1a7a48698251 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.339344] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5228b3ea-6555-7a2b-b2b9-4e6dab7767bd, 'name': SearchDatastore_Task, 'duration_secs': 0.024022} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.346898] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a50accda-8d5c-4155-b8c0-3ecb574c359a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.354223] env[61995]: DEBUG oslo_vmware.api [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 994.354223] env[61995]: value = "task-795046" [ 994.354223] env[61995]: _type = "Task" [ 994.354223] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.358390] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795045, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.387736} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.364840] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.365113] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.365279] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 994.368463] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 994.368463] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525ffcce-1206-ae8a-c1b5-060a671245bf" [ 994.368463] env[61995]: _type = "Task" [ 994.368463] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.376970] env[61995]: DEBUG oslo_vmware.api [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795046, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.383316] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525ffcce-1206-ae8a-c1b5-060a671245bf, 'name': SearchDatastore_Task, 'duration_secs': 0.023667} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.383542] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.383806] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 349667a3-6311-4c9d-bad2-ef4a5d1c52c8/349667a3-6311-4c9d-bad2-ef4a5d1c52c8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 994.384081] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e652e6a-91ae-465f-99ab-45b3c3f0c2d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.394022] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 994.394022] env[61995]: value = "task-795047" [ 994.394022] env[61995]: _type = "Task" [ 994.394022] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.402066] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.564355] env[61995]: DEBUG nova.network.neutron [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Updated VIF entry in instance network info cache for port 70414355-248a-4beb-b64e-aedd609cd31b. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 994.564747] env[61995]: DEBUG nova.network.neutron [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Updating instance_info_cache with network_info: [{"id": "70414355-248a-4beb-b64e-aedd609cd31b", "address": "fa:16:3e:e6:c3:41", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70414355-24", "ovs_interfaceid": "70414355-248a-4beb-b64e-aedd609cd31b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.590255] env[61995]: DEBUG nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 994.687585] env[61995]: INFO nova.compute.manager [-] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Took 1.96 seconds to deallocate network for instance. [ 994.852948] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.853236] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.853452] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.853640] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.853818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.855909] env[61995]: INFO nova.compute.manager [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Terminating instance [ 994.857879] env[61995]: DEBUG nova.compute.manager [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 994.858131] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 994.859017] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878a30e0-f2a5-462d-8983-b071466e3b98 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.876177] env[61995]: DEBUG oslo_vmware.api [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795046, 'name': ReconfigVM_Task, 'duration_secs': 0.272479} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.878025] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 994.882903] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.883234] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fb476f7-7e00-457d-a99d-e9bd2ca64297 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.894135] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19f40aaf-6a2e-4132-bcbd-a3c72a0aed91 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.909161] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795047, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.910821] env[61995]: DEBUG oslo_vmware.api [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 994.910821] env[61995]: value = "task-795048" [ 994.910821] env[61995]: _type = "Task" [ 994.910821] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.912630] env[61995]: DEBUG oslo_vmware.api [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 994.912630] env[61995]: value = "task-795049" [ 994.912630] env[61995]: _type = "Task" [ 994.912630] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.929339] env[61995]: DEBUG oslo_vmware.api [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795049, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.932853] env[61995]: DEBUG oslo_vmware.api [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795048, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.067639] env[61995]: DEBUG oslo_concurrency.lockutils [req-e5fffdc2-4bbc-4f1b-9963-4a97120f898c req-8eee22aa-c09d-4db6-b087-37e9e22c152a service nova] Releasing lock "refresh_cache-349667a3-6311-4c9d-bad2-ef4a5d1c52c8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.120729] env[61995]: DEBUG nova.compute.manager [req-7036d609-fd21-4690-a3f9-f50d97f6694a req-9f1c8e07-86e5-4fc4-abe4-ff2941d62f8f service nova] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Received event network-vif-deleted-fd56d4b0-9fa3-445a-a99c-67842d5494e7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 995.122389] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.122829] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.125253] env[61995]: INFO nova.compute.claims [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.201395] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.403584] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.403867] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.404041] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.404237] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.404502] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.404585] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.404980] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.405072] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.405340] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.405591] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.405850] env[61995]: DEBUG nova.virt.hardware [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.406731] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a1e6fc-791c-4c3d-b1b6-8ba3a34f8bf1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.412589] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795047, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.814705} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.415900] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 349667a3-6311-4c9d-bad2-ef4a5d1c52c8/349667a3-6311-4c9d-bad2-ef4a5d1c52c8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 995.416221] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.418476] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b468644-5272-4430-9b87-e033bd6fe14e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.424877] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b759354f-a154-4671-9f5a-1bb73efd07c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.432268] env[61995]: DEBUG oslo_vmware.api [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795048, 'name': PowerOffVM_Task, 'duration_secs': 0.261029} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.433185] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 995.433374] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 995.433898] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c48678b-ca7a-4c9d-b399-7b51ec13f2d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.438769] env[61995]: DEBUG oslo_vmware.api [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795049, 'name': ReconfigVM_Task, 'duration_secs': 0.233627} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.447600] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185367', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'name': 'volume-908f434d-3808-42d4-a891-104337fb5452', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e835ac53-29af-4bd0-b186-5c6270ccf760', 'attached_at': '', 'detached_at': '', 'volume_id': '908f434d-3808-42d4-a891-104337fb5452', 'serial': '908f434d-3808-42d4-a891-104337fb5452'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 995.450194] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 995.450194] env[61995]: value = "task-795050" [ 995.450194] env[61995]: _type = "Task" [ 995.450194] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.450674] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:c1:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ada35c98-01a9-4352-98e4-1d20ba31f928', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0bd3d577-8a66-4cce-bc58-1d82dd64bbf2', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.458148] env[61995]: DEBUG oslo.service.loopingcall [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.458726] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.461922] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9a3c4f9-acf4-4e6d-9b5d-732148877270 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.486858] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795050, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.488765] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.488765] env[61995]: value = "task-795052" [ 995.488765] env[61995]: _type = "Task" [ 995.488765] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.498192] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795052, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.555477] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 995.555734] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 995.555946] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleting the datastore file [datastore1] bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.556364] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74f0cf2b-42ab-4d82-bc2c-e671bcc40aa9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.565847] env[61995]: DEBUG oslo_vmware.api [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 995.565847] env[61995]: value = "task-795053" [ 995.565847] env[61995]: _type = "Task" [ 995.565847] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.575498] env[61995]: DEBUG oslo_vmware.api [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795053, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.598409] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "229d966c-b70d-4a51-a176-2e88488d5c4b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.598753] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.598990] env[61995]: DEBUG nova.compute.manager [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Going to confirm migration 5 {{(pid=61995) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 995.793687] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.794342] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.794473] env[61995]: DEBUG nova.objects.instance [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'flavor' on Instance uuid 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.968196] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795050, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154798} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.968534] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.969367] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b30fd3-edd7-4e47-8f75-2205c6c154cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.991777] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 349667a3-6311-4c9d-bad2-ef4a5d1c52c8/349667a3-6311-4c9d-bad2-ef4a5d1c52c8.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.992143] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd9a0ffc-10cc-43e3-b9f8-aebb3ffd866a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.007548] env[61995]: DEBUG nova.objects.instance [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.018379] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795052, 'name': CreateVM_Task, 'duration_secs': 0.466934} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.019621] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 996.019974] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 996.019974] env[61995]: value = "task-795054" [ 996.019974] env[61995]: _type = "Task" [ 996.019974] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.021230] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.021394] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.021721] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.022309] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6437662-4030-49e1-9796-bedac9aa6ef8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.031257] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 996.031257] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527a2c5f-59e2-35d5-44a9-356cd3cbae2a" [ 996.031257] env[61995]: _type = "Task" [ 996.031257] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.034720] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795054, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.043781] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527a2c5f-59e2-35d5-44a9-356cd3cbae2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.076712] env[61995]: DEBUG oslo_vmware.api [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795053, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.206291] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.206510] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.206722] env[61995]: DEBUG nova.network.neutron [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.206914] env[61995]: DEBUG nova.objects.instance [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'info_cache' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.392562] env[61995]: DEBUG nova.objects.instance [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'pci_requests' on Instance uuid 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.394911] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f3f911-5151-487d-8e3b-f75793ad989f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.403331] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f9e8a8-7d69-4628-af99-696967ba2754 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.434390] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ebec02-70b3-42a2-bf07-b19d5c17b727 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.442247] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72847414-786b-427b-8445-158b243a0492 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.455526] env[61995]: DEBUG nova.compute.provider_tree [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.533442] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795054, 'name': ReconfigVM_Task, 'duration_secs': 0.290114} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.533750] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 349667a3-6311-4c9d-bad2-ef4a5d1c52c8/349667a3-6311-4c9d-bad2-ef4a5d1c52c8.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.534372] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27dddab9-1448-4cf6-ab80-d9cdb6312264 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.545877] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527a2c5f-59e2-35d5-44a9-356cd3cbae2a, 'name': SearchDatastore_Task, 'duration_secs': 0.030071} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.547204] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.547449] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.547688] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.547836] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.548027] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.548376] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 996.548376] env[61995]: value = "task-795055" [ 996.548376] env[61995]: _type = "Task" [ 996.548376] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.548830] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3464375-3c55-4739-a6b8-5544c8d33700 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.559025] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795055, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.561526] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.561708] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 996.562420] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-793c30eb-e06d-4f74-afa0-7e16d65b6569 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.567537] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 996.567537] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52067340-2018-4ffa-64e1-a18def81a045" [ 996.567537] env[61995]: _type = "Task" [ 996.567537] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.578735] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52067340-2018-4ffa-64e1-a18def81a045, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.581680] env[61995]: DEBUG oslo_vmware.api [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795053, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.55281} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.581926] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.582125] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 996.582309] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 996.582485] env[61995]: INFO nova.compute.manager [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Took 1.72 seconds to destroy the instance on the hypervisor. [ 996.582725] env[61995]: DEBUG oslo.service.loopingcall [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.582911] env[61995]: DEBUG nova.compute.manager [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.583016] env[61995]: DEBUG nova.network.neutron [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 996.676166] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.899229] env[61995]: DEBUG nova.objects.base [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Object Instance<453f46c7-5bad-4ca2-b228-f76e62fbd03e> lazy-loaded attributes: flavor,pci_requests {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 996.900109] env[61995]: DEBUG nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 996.959516] env[61995]: DEBUG nova.scheduler.client.report [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.986689] env[61995]: DEBUG nova.policy [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 997.018267] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c0fb4ffc-087e-4185-9c4e-f573a7d327e3 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.348s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.019443] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.343s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.019645] env[61995]: DEBUG nova.compute.manager [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.020520] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31336b95-a608-44db-b488-2bf830af0e8b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.028322] env[61995]: DEBUG nova.compute.manager [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 997.028912] env[61995]: DEBUG nova.objects.instance [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.062734] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795055, 'name': Rename_Task, 'duration_secs': 0.153507} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.063015] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 997.063268] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-815e4e13-682c-4fad-8d1e-853ce5f12014 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.071368] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 997.071368] env[61995]: value = "task-795056" [ 997.071368] env[61995]: _type = "Task" [ 997.071368] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.083080] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52067340-2018-4ffa-64e1-a18def81a045, 'name': SearchDatastore_Task, 'duration_secs': 0.013052} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.086819] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795056, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.087047] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db2d51de-fd75-449f-a026-ad1b615ac572 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.093579] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 997.093579] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52fa2ee9-5831-eca1-2a96-c23a9b656ee2" [ 997.093579] env[61995]: _type = "Task" [ 997.093579] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.103547] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fa2ee9-5831-eca1-2a96-c23a9b656ee2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.460029] env[61995]: DEBUG nova.compute.manager [req-11d0590e-384f-40c4-8056-c6f7b6b84802 req-67e382f2-1b13-4e3e-a4bc-0e9a5f2b2e3e service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Received event network-vif-deleted-a0291e27-4d18-41af-ba79-3d058998150e {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 997.460196] env[61995]: INFO nova.compute.manager [req-11d0590e-384f-40c4-8056-c6f7b6b84802 req-67e382f2-1b13-4e3e-a4bc-0e9a5f2b2e3e service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Neutron deleted interface a0291e27-4d18-41af-ba79-3d058998150e; detaching it from the instance and deleting it from the info cache [ 997.460870] env[61995]: DEBUG nova.network.neutron [req-11d0590e-384f-40c4-8056-c6f7b6b84802 req-67e382f2-1b13-4e3e-a4bc-0e9a5f2b2e3e service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.467452] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.467939] env[61995]: DEBUG nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 997.475783] env[61995]: DEBUG nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Successfully created port: 8c9dcf57-6289-4787-94d0-0d19f3a76395 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.478147] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.277s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.478487] env[61995]: DEBUG nova.objects.instance [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lazy-loading 'resources' on Instance uuid 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.480329] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.480588] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.489172] env[61995]: DEBUG nova.network.neutron [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.536829] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 997.537199] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b9245c1-670e-4b13-9768-f0dfe6f58763 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.545982] env[61995]: DEBUG oslo_vmware.api [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 997.545982] env[61995]: value = "task-795057" [ 997.545982] env[61995]: _type = "Task" [ 997.545982] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.555858] env[61995]: DEBUG oslo_vmware.api [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.586012] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795056, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.604396] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fa2ee9-5831-eca1-2a96-c23a9b656ee2, 'name': SearchDatastore_Task, 'duration_secs': 0.015835} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.604897] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.605821] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 997.606214] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbac8da2-8d57-4024-aaa1-b51751928d01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.615744] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 997.615744] env[61995]: value = "task-795058" [ 997.615744] env[61995]: _type = "Task" [ 997.615744] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.626498] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.835748] env[61995]: DEBUG nova.network.neutron [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.975788] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf0b6b58-69aa-4ceb-9a57-2da7a2b732e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.979924] env[61995]: DEBUG nova.compute.utils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.981898] env[61995]: DEBUG nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.982088] env[61995]: DEBUG nova.network.neutron [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 997.987223] env[61995]: DEBUG nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 997.997936] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36aafeff-d781-4c79-bba0-05866b78488d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.010749] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.011057] env[61995]: DEBUG nova.objects.instance [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'migration_context' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.036399] env[61995]: DEBUG nova.compute.manager [req-11d0590e-384f-40c4-8056-c6f7b6b84802 req-67e382f2-1b13-4e3e-a4bc-0e9a5f2b2e3e service nova] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Detach interface failed, port_id=a0291e27-4d18-41af-ba79-3d058998150e, reason: Instance bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 998.042428] env[61995]: DEBUG nova.policy [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '671c956f9796411e9b66268420075628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26c7cc2bdcb6462d8154127098e94875', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 998.058999] env[61995]: DEBUG oslo_vmware.api [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795057, 'name': PowerOffVM_Task, 'duration_secs': 0.264221} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.058999] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 998.058999] env[61995]: DEBUG nova.compute.manager [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.058999] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb28d22-7e90-4e71-aa88-99e505b54922 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.088310] env[61995]: DEBUG oslo_vmware.api [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795056, 'name': PowerOnVM_Task, 'duration_secs': 0.648196} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.088310] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.088310] env[61995]: INFO nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Took 7.95 seconds to spawn the instance on the hypervisor. [ 998.088310] env[61995]: DEBUG nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.088635] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188ddc15-a40b-49fd-9f6f-8674d6a439a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.133410] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795058, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.276655] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102fb2d8-1087-4aa7-8274-3666cdec80fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.285850] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d8fabd-d0da-4ae7-88f6-f70fb4969468 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.322151] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775c4f12-3caf-4b42-a9ff-3e2cb600c8c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.331533] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb952b22-4f7a-4155-b20a-81785dc878ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.340109] env[61995]: INFO nova.compute.manager [-] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Took 1.76 seconds to deallocate network for instance. [ 998.351784] env[61995]: DEBUG nova.compute.provider_tree [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.354305] env[61995]: DEBUG nova.network.neutron [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Successfully created port: 41529e93-5bd1-44e2-9904-dc6702ff0282 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.487095] env[61995]: DEBUG nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.507814] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.513757] env[61995]: DEBUG nova.objects.base [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Object Instance<229d966c-b70d-4a51-a176-2e88488d5c4b> lazy-loaded attributes: info_cache,migration_context {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 998.516652] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84b5e87-97ff-462b-8826-d5b35ed52ccd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.534890] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9575dd5-5f3f-41cc-ad93-7f965566a968 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.541561] env[61995]: DEBUG oslo_vmware.api [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 998.541561] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526c948c-0f45-1d58-8ad2-0c5539cb6fe5" [ 998.541561] env[61995]: _type = "Task" [ 998.541561] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.550121] env[61995]: DEBUG oslo_vmware.api [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526c948c-0f45-1d58-8ad2-0c5539cb6fe5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.574472] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e07c4f8f-160e-4941-b26c-65d920e6e7ff tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.611623] env[61995]: INFO nova.compute.manager [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Took 14.50 seconds to build instance. [ 998.626401] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795058, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.79489} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.627249] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 998.627454] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.627711] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64f98db2-72a6-4365-8620-320bc4fd021d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.636830] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 998.636830] env[61995]: value = "task-795059" [ 998.636830] env[61995]: _type = "Task" [ 998.636830] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.646075] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795059, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.860886] env[61995]: DEBUG nova.scheduler.client.report [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.864972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.053107] env[61995]: DEBUG oslo_vmware.api [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526c948c-0f45-1d58-8ad2-0c5539cb6fe5, 'name': SearchDatastore_Task, 'duration_secs': 0.04125} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.053413] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.114277] env[61995]: DEBUG oslo_concurrency.lockutils [None req-674bfa65-62ea-4b41-b18a-3228fb799911 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.013s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.145568] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795059, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07263} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.145859] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.146654] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7ecd2c-607f-4e04-88e8-5845daed91b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.169909] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.170204] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42ac07be-a1df-4ff9-97ee-5fa0393f6ea5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.185362] env[61995]: DEBUG nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Successfully updated port: 8c9dcf57-6289-4787-94d0-0d19f3a76395 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.192402] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 999.192402] env[61995]: value = "task-795060" [ 999.192402] env[61995]: _type = "Task" [ 999.192402] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.202238] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795060, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.276907] env[61995]: DEBUG nova.objects.instance [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.368198] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.889s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.370207] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.862s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.371617] env[61995]: INFO nova.compute.claims [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.390272] env[61995]: INFO nova.scheduler.client.report [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Deleted allocations for instance 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4 [ 999.494189] env[61995]: DEBUG nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 999.520199] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.520462] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.520624] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.520809] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.520956] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.521116] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.521330] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.521492] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.521663] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.521827] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.522012] env[61995]: DEBUG nova.virt.hardware [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.522899] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38160097-9827-4f3b-8ef5-2d467be644ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.532085] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d925d966-b42e-426e-9bc4-ae30797a84d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.625521] env[61995]: DEBUG nova.compute.manager [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-vif-plugged-8c9dcf57-6289-4787-94d0-0d19f3a76395 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.625893] env[61995]: DEBUG oslo_concurrency.lockutils [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.626260] env[61995]: DEBUG oslo_concurrency.lockutils [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.626574] env[61995]: DEBUG oslo_concurrency.lockutils [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.626880] env[61995]: DEBUG nova.compute.manager [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] No waiting events found dispatching network-vif-plugged-8c9dcf57-6289-4787-94d0-0d19f3a76395 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 999.627208] env[61995]: WARNING nova.compute.manager [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received unexpected event network-vif-plugged-8c9dcf57-6289-4787-94d0-0d19f3a76395 for instance with vm_state active and task_state None. [ 999.627565] env[61995]: DEBUG nova.compute.manager [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-changed-8c9dcf57-6289-4787-94d0-0d19f3a76395 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.627874] env[61995]: DEBUG nova.compute.manager [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing instance network info cache due to event network-changed-8c9dcf57-6289-4787-94d0-0d19f3a76395. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 999.628276] env[61995]: DEBUG oslo_concurrency.lockutils [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.628547] env[61995]: DEBUG oslo_concurrency.lockutils [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.628846] env[61995]: DEBUG nova.network.neutron [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing network info cache for port 8c9dcf57-6289-4787-94d0-0d19f3a76395 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 999.688289] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.703735] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.782345] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.782470] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.783030] env[61995]: DEBUG nova.network.neutron [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.783030] env[61995]: DEBUG nova.objects.instance [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'info_cache' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.837271] env[61995]: DEBUG nova.network.neutron [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Successfully updated port: 41529e93-5bd1-44e2-9904-dc6702ff0282 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.871762] env[61995]: DEBUG nova.compute.manager [req-3a56e847-9e3e-4a62-bb40-91b3f4014353 req-9ffadad2-b3e2-4e34-a3fd-fb4dce69feb0 service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Received event network-vif-plugged-41529e93-5bd1-44e2-9904-dc6702ff0282 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.872022] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a56e847-9e3e-4a62-bb40-91b3f4014353 req-9ffadad2-b3e2-4e34-a3fd-fb4dce69feb0 service nova] Acquiring lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.872285] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a56e847-9e3e-4a62-bb40-91b3f4014353 req-9ffadad2-b3e2-4e34-a3fd-fb4dce69feb0 service nova] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.872436] env[61995]: DEBUG oslo_concurrency.lockutils [req-3a56e847-9e3e-4a62-bb40-91b3f4014353 req-9ffadad2-b3e2-4e34-a3fd-fb4dce69feb0 service nova] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.872624] env[61995]: DEBUG nova.compute.manager [req-3a56e847-9e3e-4a62-bb40-91b3f4014353 req-9ffadad2-b3e2-4e34-a3fd-fb4dce69feb0 service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] No waiting events found dispatching network-vif-plugged-41529e93-5bd1-44e2-9904-dc6702ff0282 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 999.872804] env[61995]: WARNING nova.compute.manager [req-3a56e847-9e3e-4a62-bb40-91b3f4014353 req-9ffadad2-b3e2-4e34-a3fd-fb4dce69feb0 service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Received unexpected event network-vif-plugged-41529e93-5bd1-44e2-9904-dc6702ff0282 for instance with vm_state building and task_state spawning. [ 999.897120] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5e3f9aac-d146-4921-a1a4-801a73d5c0ab tempest-ServerDiskConfigTestJSON-302716645 tempest-ServerDiskConfigTestJSON-302716645-project-member] Lock "3a2adf56-8071-476d-8cf8-5f6ece6a1ac4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.562s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.206055] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795060, 'name': ReconfigVM_Task, 'duration_secs': 0.778351} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.206430] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 9627f278-73d6-45d3-b60d-776e373eef73/9627f278-73d6-45d3-b60d-776e373eef73.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.207223] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6b0ea5e-b951-4ba6-9f41-aa67dd60a9cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.216521] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1000.216521] env[61995]: value = "task-795061" [ 1000.216521] env[61995]: _type = "Task" [ 1000.216521] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.229635] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795061, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.286493] env[61995]: DEBUG nova.objects.base [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1000.339765] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "refresh_cache-9e30eadd-5694-4fc7-8b54-2cf1d1571504" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.339972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "refresh_cache-9e30eadd-5694-4fc7-8b54-2cf1d1571504" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.340150] env[61995]: DEBUG nova.network.neutron [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.376573] env[61995]: DEBUG nova.network.neutron [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Added VIF to instance network info cache for port 8c9dcf57-6289-4787-94d0-0d19f3a76395. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 1000.377146] env[61995]: DEBUG nova.network.neutron [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "address": "fa:16:3e:c5:f2:9f", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c9dcf57-62", "ovs_interfaceid": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.435273] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.435587] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.435810] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.440922] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.440922] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.003s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.442377] env[61995]: INFO nova.compute.manager [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Terminating instance [ 1000.444503] env[61995]: DEBUG nova.compute.manager [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.445359] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1000.445622] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42159459-7549-455f-91ef-db7e2ae9be00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.454182] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.458024] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7a830f0-bf7f-4abf-8194-8f8ddfd916e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.465826] env[61995]: DEBUG oslo_vmware.api [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1000.465826] env[61995]: value = "task-795062" [ 1000.465826] env[61995]: _type = "Task" [ 1000.465826] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.474647] env[61995]: DEBUG oslo_vmware.api [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.618043] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f025b18a-5c12-46e0-bcf9-642ef73461d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.629501] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2028a5-72da-40b6-b1d1-9a94721770c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.665194] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c4080a-c731-4912-8aa0-d933a0a9f3ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.675184] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19b3263-ea4e-4d8b-82c7-69ef28d72517 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.690906] env[61995]: DEBUG nova.compute.provider_tree [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.727451] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795061, 'name': Rename_Task, 'duration_secs': 0.161671} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.727741] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1000.727981] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f02faebf-4a9c-4ae3-a7d4-2e694af0ea6f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.736825] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1000.736825] env[61995]: value = "task-795063" [ 1000.736825] env[61995]: _type = "Task" [ 1000.736825] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.753219] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795063, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.874363] env[61995]: DEBUG nova.network.neutron [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1000.882390] env[61995]: DEBUG oslo_concurrency.lockutils [req-ae967e39-106e-483a-b5e6-9478c2b00986 req-ab08fbf8-979d-4183-ba86-0cca4b41477d service nova] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.882834] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.883036] env[61995]: DEBUG nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.976388] env[61995]: DEBUG oslo_vmware.api [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795062, 'name': PowerOffVM_Task, 'duration_secs': 0.241793} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.978907] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1000.979100] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1000.979361] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7001f0ed-9581-4bbe-bfd0-45a46040ccce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.058866] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1001.059166] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1001.059379] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleting the datastore file [datastore2] 349667a3-6311-4c9d-bad2-ef4a5d1c52c8 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.060393] env[61995]: DEBUG nova.network.neutron [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Updating instance_info_cache with network_info: [{"id": "41529e93-5bd1-44e2-9904-dc6702ff0282", "address": "fa:16:3e:60:9f:aa", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41529e93-5b", "ovs_interfaceid": "41529e93-5bd1-44e2-9904-dc6702ff0282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.061677] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc1b7f81-0df7-4df7-895a-34a7c4739ec6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.071399] env[61995]: DEBUG oslo_vmware.api [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1001.071399] env[61995]: value = "task-795065" [ 1001.071399] env[61995]: _type = "Task" [ 1001.071399] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.082266] env[61995]: DEBUG oslo_vmware.api [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795065, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.123622] env[61995]: DEBUG nova.network.neutron [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating instance_info_cache with network_info: [{"id": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "address": "fa:16:3e:c6:74:e1", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c280b5d-bf", "ovs_interfaceid": "8c280b5d-bf84-4a8d-8ab7-4230645c59d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.195146] env[61995]: DEBUG nova.scheduler.client.report [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.249638] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795063, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.430469] env[61995]: WARNING nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 1001.430738] env[61995]: WARNING nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 1001.430936] env[61995]: WARNING nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] 8c9dcf57-6289-4787-94d0-0d19f3a76395 already exists in list: port_ids containing: ['8c9dcf57-6289-4787-94d0-0d19f3a76395']. ignoring it [ 1001.566354] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "refresh_cache-9e30eadd-5694-4fc7-8b54-2cf1d1571504" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.566675] env[61995]: DEBUG nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Instance network_info: |[{"id": "41529e93-5bd1-44e2-9904-dc6702ff0282", "address": "fa:16:3e:60:9f:aa", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41529e93-5b", "ovs_interfaceid": "41529e93-5bd1-44e2-9904-dc6702ff0282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.567119] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:9f:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41529e93-5bd1-44e2-9904-dc6702ff0282', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.574909] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating folder: Project (26c7cc2bdcb6462d8154127098e94875). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1001.575068] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a8f69d4-098b-4aa6-83a3-1710bf218c09 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.585609] env[61995]: DEBUG oslo_vmware.api [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795065, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18779} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.585843] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.586176] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1001.586421] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1001.586610] env[61995]: INFO nova.compute.manager [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1001.586860] env[61995]: DEBUG oslo.service.loopingcall [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.587100] env[61995]: DEBUG nova.compute.manager [-] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.587327] env[61995]: DEBUG nova.network.neutron [-] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1001.590682] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created folder: Project (26c7cc2bdcb6462d8154127098e94875) in parent group-v185203. [ 1001.590930] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating folder: Instances. Parent ref: group-v185387. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1001.591510] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e041a06-ffa0-40d7-a9df-ef44911f08ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.603033] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created folder: Instances in parent group-v185387. [ 1001.603033] env[61995]: DEBUG oslo.service.loopingcall [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.603208] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1001.603307] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7827568-823d-4b5a-bcc2-f5d21c4952ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.623917] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.623917] env[61995]: value = "task-795068" [ 1001.623917] env[61995]: _type = "Task" [ 1001.623917] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.627293] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "refresh_cache-e835ac53-29af-4bd0-b186-5c6270ccf760" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.633850] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795068, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.700553] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.701155] env[61995]: DEBUG nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1001.705768] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.841s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.706031] env[61995]: DEBUG nova.objects.instance [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'resources' on Instance uuid bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.747606] env[61995]: DEBUG oslo_vmware.api [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795063, 'name': PowerOnVM_Task, 'duration_secs': 0.902622} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.747889] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.748147] env[61995]: DEBUG nova.compute.manager [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.749199] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7820f695-8169-4c59-ade0-c5d61bdcec76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.950346] env[61995]: DEBUG nova.network.neutron [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "address": "fa:16:3e:c5:f2:9f", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c9dcf57-62", "ovs_interfaceid": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.110049] env[61995]: DEBUG nova.compute.manager [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Received event network-changed-41529e93-5bd1-44e2-9904-dc6702ff0282 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.110429] env[61995]: DEBUG nova.compute.manager [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Refreshing instance network info cache due to event network-changed-41529e93-5bd1-44e2-9904-dc6702ff0282. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1002.110695] env[61995]: DEBUG oslo_concurrency.lockutils [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] Acquiring lock "refresh_cache-9e30eadd-5694-4fc7-8b54-2cf1d1571504" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.110848] env[61995]: DEBUG oslo_concurrency.lockutils [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] Acquired lock "refresh_cache-9e30eadd-5694-4fc7-8b54-2cf1d1571504" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.111026] env[61995]: DEBUG nova.network.neutron [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Refreshing network info cache for port 41529e93-5bd1-44e2-9904-dc6702ff0282 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1002.130328] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1002.130663] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a8ad0d3-19f3-46c7-b652-e3ec434dcba1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.135888] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795068, 'name': CreateVM_Task, 'duration_secs': 0.370418} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.136439] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1002.137214] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.137463] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.137847] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.138541] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-476d8eee-81d3-46af-9f07-a3317196b24a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.142286] env[61995]: DEBUG oslo_vmware.api [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1002.142286] env[61995]: value = "task-795069" [ 1002.142286] env[61995]: _type = "Task" [ 1002.142286] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.152248] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1002.152248] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5208c00f-462e-91e9-80a5-e276f63057b7" [ 1002.152248] env[61995]: _type = "Task" [ 1002.152248] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.162029] env[61995]: DEBUG oslo_vmware.api [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795069, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.167129] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5208c00f-462e-91e9-80a5-e276f63057b7, 'name': SearchDatastore_Task, 'duration_secs': 0.011783} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.167433] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.167671] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.167909] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.168117] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.168313] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.168579] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-521dd872-eb6c-42f0-9554-4474e457a499 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.186021] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.186021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1002.186021] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca552663-f6f2-4067-8370-878053bf5efc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.191414] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1002.191414] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e82307-16ca-ac15-eaaa-9bec6b504fed" [ 1002.191414] env[61995]: _type = "Task" [ 1002.191414] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.202651] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e82307-16ca-ac15-eaaa-9bec6b504fed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.209182] env[61995]: DEBUG nova.compute.utils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1002.210628] env[61995]: DEBUG nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1002.210800] env[61995]: DEBUG nova.network.neutron [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1002.247123] env[61995]: DEBUG nova.compute.manager [req-80a653ac-455e-414f-8c92-184b6d1f2fd1 req-ca39df58-fbd6-4733-acd5-ba513745fa57 service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Received event network-vif-deleted-70414355-248a-4beb-b64e-aedd609cd31b {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.247402] env[61995]: INFO nova.compute.manager [req-80a653ac-455e-414f-8c92-184b6d1f2fd1 req-ca39df58-fbd6-4733-acd5-ba513745fa57 service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Neutron deleted interface 70414355-248a-4beb-b64e-aedd609cd31b; detaching it from the instance and deleting it from the info cache [ 1002.247623] env[61995]: DEBUG nova.network.neutron [req-80a653ac-455e-414f-8c92-184b6d1f2fd1 req-ca39df58-fbd6-4733-acd5-ba513745fa57 service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.259048] env[61995]: DEBUG nova.policy [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0acd3ba123ae4a7da49803fe40ced8d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f543b47c4254f4ebeca453aea3123c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1002.273252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.454522] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.454679] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.454748] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.455803] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73bd41a0-d404-478c-8ad6-67434d41ad4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.476148] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.476407] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.476574] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.476759] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.476910] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.477076] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.477417] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.477482] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.477641] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.477795] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.478120] env[61995]: DEBUG nova.virt.hardware [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.485144] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfiguring VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1002.488174] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2dee741-c78b-4fd1-99d6-eabd992b1f41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.515256] env[61995]: DEBUG oslo_vmware.api [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1002.515256] env[61995]: value = "task-795070" [ 1002.515256] env[61995]: _type = "Task" [ 1002.515256] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.527863] env[61995]: DEBUG oslo_vmware.api [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795070, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.530987] env[61995]: DEBUG nova.network.neutron [-] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.541448] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e3954d-4b26-48b9-b2ff-c5a3fb38a9c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.555884] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c42da1-8cc3-4978-9221-059be853fbdb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.595497] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49be9def-39f4-4a17-8ef0-22f8227c5fef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.600036] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f8a733-5f5b-4325-bb15-0ad75fc05c0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.616956] env[61995]: DEBUG nova.compute.provider_tree [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.653017] env[61995]: DEBUG oslo_vmware.api [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795069, 'name': PowerOnVM_Task, 'duration_secs': 0.420821} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.653422] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1002.653742] env[61995]: DEBUG nova.compute.manager [None req-ad983a85-a434-4ca2-bfc8-035fe93f54f1 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.654513] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7584470-92ff-4735-9ce4-b4f0d7aa32bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.706173] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e82307-16ca-ac15-eaaa-9bec6b504fed, 'name': SearchDatastore_Task, 'duration_secs': 0.012838} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.706173] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed85ea2a-6991-4045-97d0-ec52348a6529 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.714127] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1002.714127] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52139d91-572d-48e3-39e0-152f077984a4" [ 1002.714127] env[61995]: _type = "Task" [ 1002.714127] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.718253] env[61995]: DEBUG nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1002.731635] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52139d91-572d-48e3-39e0-152f077984a4, 'name': SearchDatastore_Task, 'duration_secs': 0.010744} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.732546] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.732732] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 9e30eadd-5694-4fc7-8b54-2cf1d1571504/9e30eadd-5694-4fc7-8b54-2cf1d1571504.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1002.733092] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2620b0a4-c3ee-4343-9b3c-f3b4a34a4e35 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.741269] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1002.741269] env[61995]: value = "task-795071" [ 1002.741269] env[61995]: _type = "Task" [ 1002.741269] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.750591] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795071, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.753387] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ff2824b-6f5d-4772-9112-28fa8b14e5e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.762472] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62508f02-58c4-4e6e-907e-b1031de77e40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.774286] env[61995]: DEBUG nova.network.neutron [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Successfully created port: ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.799562] env[61995]: DEBUG nova.compute.manager [req-80a653ac-455e-414f-8c92-184b6d1f2fd1 req-ca39df58-fbd6-4733-acd5-ba513745fa57 service nova] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Detach interface failed, port_id=70414355-248a-4beb-b64e-aedd609cd31b, reason: Instance 349667a3-6311-4c9d-bad2-ef4a5d1c52c8 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1002.982749] env[61995]: DEBUG nova.network.neutron [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Updated VIF entry in instance network info cache for port 41529e93-5bd1-44e2-9904-dc6702ff0282. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1002.983186] env[61995]: DEBUG nova.network.neutron [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Updating instance_info_cache with network_info: [{"id": "41529e93-5bd1-44e2-9904-dc6702ff0282", "address": "fa:16:3e:60:9f:aa", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41529e93-5b", "ovs_interfaceid": "41529e93-5bd1-44e2-9904-dc6702ff0282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.032351] env[61995]: DEBUG oslo_vmware.api [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795070, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.035520] env[61995]: INFO nova.compute.manager [-] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Took 1.45 seconds to deallocate network for instance. [ 1003.123840] env[61995]: DEBUG nova.scheduler.client.report [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.251355] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795071, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.260568] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.261039] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.261039] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1003.486361] env[61995]: DEBUG oslo_concurrency.lockutils [req-21dab4ed-449b-4a22-8f35-42dc8627453e req-e186c075-1802-4966-ad80-5dea5be0273a service nova] Releasing lock "refresh_cache-9e30eadd-5694-4fc7-8b54-2cf1d1571504" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.526301] env[61995]: DEBUG oslo_vmware.api [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795070, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.543719] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.631055] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.632522] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.579s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.669890] env[61995]: INFO nova.scheduler.client.report [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted allocations for instance bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31 [ 1003.739661] env[61995]: DEBUG nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1003.752268] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795071, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513491} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.752596] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 9e30eadd-5694-4fc7-8b54-2cf1d1571504/9e30eadd-5694-4fc7-8b54-2cf1d1571504.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1003.752762] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.753078] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8e4cac0-2c39-4639-816a-4f8528060d4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.761348] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1003.761348] env[61995]: value = "task-795072" [ 1003.761348] env[61995]: _type = "Task" [ 1003.761348] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.764682] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.764921] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.765099] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.765295] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.765464] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.765620] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.765860] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.766051] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.766228] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.766414] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.766594] env[61995]: DEBUG nova.virt.hardware [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.768619] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febf1df3-203e-409b-9ead-8e7fc486a9a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.776059] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795072, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.779154] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b361b16c-61a3-4d07-9e67-6c587fb91d50 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.031282] env[61995]: DEBUG oslo_vmware.api [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795070, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.179267] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f56b6e6a-42a5-4b80-81f8-6374e3516260 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.325s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.277400] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795072, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069421} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.278014] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.281330] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de14b54-08fa-4201-8e28-4fb905aaf641 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.307339] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 9e30eadd-5694-4fc7-8b54-2cf1d1571504/9e30eadd-5694-4fc7-8b54-2cf1d1571504.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.311030] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5c14d1b-e322-4372-ada8-ac0d931adf81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.333166] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1004.333166] env[61995]: value = "task-795073" [ 1004.333166] env[61995]: _type = "Task" [ 1004.333166] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.345705] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795073, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.437021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05ca330-2511-4394-bbfc-444201d69b84 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.447878] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0a945a-5270-444e-8e74-99293d897532 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.500577] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea58f91e-2a55-466c-8843-2a89d6b870cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.519738] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80da60ba-5d25-48be-97a5-7dbf06d8cfbc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.541914] env[61995]: DEBUG nova.network.neutron [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Successfully updated port: ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.543428] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "9627f278-73d6-45d3-b60d-776e373eef73" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.543526] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "9627f278-73d6-45d3-b60d-776e373eef73" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.543729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "9627f278-73d6-45d3-b60d-776e373eef73-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.543921] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "9627f278-73d6-45d3-b60d-776e373eef73-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.544154] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "9627f278-73d6-45d3-b60d-776e373eef73-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.546496] env[61995]: DEBUG oslo_vmware.api [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795070, 'name': ReconfigVM_Task, 'duration_secs': 1.927718} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.546496] env[61995]: DEBUG nova.compute.provider_tree [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.547536] env[61995]: INFO nova.compute.manager [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Terminating instance [ 1004.549501] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.549707] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfigured VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1004.552501] env[61995]: DEBUG nova.compute.manager [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.552717] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.553585] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ec571f-94a7-4780-94cb-2ef352e7da0b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.562861] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.563248] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d871aba-8a01-426b-bd6a-cf7f2755d11c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.570738] env[61995]: DEBUG oslo_vmware.api [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1004.570738] env[61995]: value = "task-795074" [ 1004.570738] env[61995]: _type = "Task" [ 1004.570738] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.579430] env[61995]: DEBUG oslo_vmware.api [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795074, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.843621] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795073, 'name': ReconfigVM_Task, 'duration_secs': 0.385548} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.843976] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 9e30eadd-5694-4fc7-8b54-2cf1d1571504/9e30eadd-5694-4fc7-8b54-2cf1d1571504.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.844733] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90579a48-fdc0-4e98-aec3-4512ac1ff336 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.851229] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1004.851229] env[61995]: value = "task-795075" [ 1004.851229] env[61995]: _type = "Task" [ 1004.851229] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.861639] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795075, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.918374] env[61995]: DEBUG nova.compute.manager [req-41b92095-b06b-4afd-af38-da98e6f15f00 req-ea43182e-a870-47d8-bb6e-756d7d4c9097 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Received event network-vif-plugged-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1004.918619] env[61995]: DEBUG oslo_concurrency.lockutils [req-41b92095-b06b-4afd-af38-da98e6f15f00 req-ea43182e-a870-47d8-bb6e-756d7d4c9097 service nova] Acquiring lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.918849] env[61995]: DEBUG oslo_concurrency.lockutils [req-41b92095-b06b-4afd-af38-da98e6f15f00 req-ea43182e-a870-47d8-bb6e-756d7d4c9097 service nova] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.919549] env[61995]: DEBUG oslo_concurrency.lockutils [req-41b92095-b06b-4afd-af38-da98e6f15f00 req-ea43182e-a870-47d8-bb6e-756d7d4c9097 service nova] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.919809] env[61995]: DEBUG nova.compute.manager [req-41b92095-b06b-4afd-af38-da98e6f15f00 req-ea43182e-a870-47d8-bb6e-756d7d4c9097 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] No waiting events found dispatching network-vif-plugged-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.919993] env[61995]: WARNING nova.compute.manager [req-41b92095-b06b-4afd-af38-da98e6f15f00 req-ea43182e-a870-47d8-bb6e-756d7d4c9097 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Received unexpected event network-vif-plugged-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d for instance with vm_state building and task_state spawning. [ 1005.048832] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.052019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.052019] env[61995]: DEBUG nova.network.neutron [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.053633] env[61995]: DEBUG nova.scheduler.client.report [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.060589] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7238fbca-e42e-4bbd-a79e-19abdc1d6398 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.264s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.081759] env[61995]: DEBUG oslo_vmware.api [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795074, 'name': PowerOffVM_Task, 'duration_secs': 0.166296} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.082133] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1005.083192] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1005.083192] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-923d56a8-0b0e-4fb0-9b16-75f47d76728b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.168339] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1005.168580] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1005.168769] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleting the datastore file [datastore2] 9627f278-73d6-45d3-b60d-776e373eef73 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.169149] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a386297f-0146-42f9-9277-8572b1ab4048 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.175408] env[61995]: DEBUG oslo_vmware.api [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1005.175408] env[61995]: value = "task-795077" [ 1005.175408] env[61995]: _type = "Task" [ 1005.175408] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.186646] env[61995]: DEBUG oslo_vmware.api [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.362791] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795075, 'name': Rename_Task, 'duration_secs': 0.153367} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.363174] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1005.363512] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66f6db50-b001-464e-8a80-743b4283d18d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.369775] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1005.369775] env[61995]: value = "task-795078" [ 1005.369775] env[61995]: _type = "Task" [ 1005.369775] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.378577] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795078, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.384538] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "06886222-5f7f-482e-b5ee-afd7326f2c70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.384885] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.600967] env[61995]: DEBUG nova.network.neutron [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1005.685829] env[61995]: DEBUG oslo_vmware.api [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176496} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.686112] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.686306] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1005.686487] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1005.686693] env[61995]: INFO nova.compute.manager [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1005.686960] env[61995]: DEBUG oslo.service.loopingcall [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.687190] env[61995]: DEBUG nova.compute.manager [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1005.687290] env[61995]: DEBUG nova.network.neutron [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1005.808376] env[61995]: DEBUG nova.network.neutron [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.880946] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795078, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.889050] env[61995]: DEBUG nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.015227] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.015538] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.073023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.440s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.073023] env[61995]: DEBUG nova.compute.manager [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61995) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1006.076040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.803s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.076248] env[61995]: DEBUG nova.objects.instance [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1006.310901] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.312209] env[61995]: DEBUG nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Instance network_info: |[{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1006.312934] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:13:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff48bd88-ebef-44df-92d3-1a4b9c6ff71d', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1006.321635] env[61995]: DEBUG oslo.service.loopingcall [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.322130] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1006.322743] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9013abad-f0e9-4124-ae02-588b2080bf72 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.347185] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1006.347185] env[61995]: value = "task-795079" [ 1006.347185] env[61995]: _type = "Task" [ 1006.347185] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.356849] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795079, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.381819] env[61995]: DEBUG oslo_vmware.api [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795078, 'name': PowerOnVM_Task, 'duration_secs': 0.791832} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.382263] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1006.382531] env[61995]: INFO nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Took 6.89 seconds to spawn the instance on the hypervisor. [ 1006.382730] env[61995]: DEBUG nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.383543] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84cf66f-c994-42f6-8511-de7b13619273 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.394151] env[61995]: DEBUG nova.compute.manager [req-b81b4d25-eef0-420a-b5e1-764bd8af6d04 req-75ea0af1-c71a-45d9-b37d-d9db1ac962f5 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Received event network-vif-deleted-0bd3d577-8a66-4cce-bc58-1d82dd64bbf2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1006.394391] env[61995]: INFO nova.compute.manager [req-b81b4d25-eef0-420a-b5e1-764bd8af6d04 req-75ea0af1-c71a-45d9-b37d-d9db1ac962f5 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Neutron deleted interface 0bd3d577-8a66-4cce-bc58-1d82dd64bbf2; detaching it from the instance and deleting it from the info cache [ 1006.394567] env[61995]: DEBUG nova.network.neutron [req-b81b4d25-eef0-420a-b5e1-764bd8af6d04 req-75ea0af1-c71a-45d9-b37d-d9db1ac962f5 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.421203] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.518190] env[61995]: DEBUG nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.612434] env[61995]: DEBUG nova.network.neutron [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.648459] env[61995]: INFO nova.scheduler.client.report [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted allocation for migration 82d397d5-8c3f-445a-ab18-e97956239d3f [ 1006.753205] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "a48d837d-1008-4248-bbe0-14321368e138" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.753205] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "a48d837d-1008-4248-bbe0-14321368e138" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.857945] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795079, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.905370] env[61995]: INFO nova.compute.manager [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Took 11.81 seconds to build instance. [ 1006.906534] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2bc96feb-1423-416c-8728-e052d3d4e27d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.916877] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ed2bca-8f31-42f0-8d97-aae2b8d62770 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.949917] env[61995]: DEBUG nova.compute.manager [req-b81b4d25-eef0-420a-b5e1-764bd8af6d04 req-75ea0af1-c71a-45d9-b37d-d9db1ac962f5 service nova] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Detach interface failed, port_id=0bd3d577-8a66-4cce-bc58-1d82dd64bbf2, reason: Instance 9627f278-73d6-45d3-b60d-776e373eef73 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1007.029370] env[61995]: DEBUG nova.compute.manager [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Received event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.029537] env[61995]: DEBUG nova.compute.manager [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing instance network info cache due to event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1007.029757] env[61995]: DEBUG oslo_concurrency.lockutils [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] Acquiring lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.029904] env[61995]: DEBUG oslo_concurrency.lockutils [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] Acquired lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.030083] env[61995]: DEBUG nova.network.neutron [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1007.041061] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.089969] env[61995]: DEBUG oslo_concurrency.lockutils [None req-238c2811-f1af-4cab-995b-3c1283c0bd7d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.091140] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.548s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.091467] env[61995]: DEBUG nova.objects.instance [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'resources' on Instance uuid 349667a3-6311-4c9d-bad2-ef4a5d1c52c8 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.115377] env[61995]: INFO nova.compute.manager [-] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Took 1.43 seconds to deallocate network for instance. [ 1007.157626] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7c91531d-7bda-4ec6-a876-3240c52af8be tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.559s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.254390] env[61995]: DEBUG nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1007.282845] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.282979] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.283599] env[61995]: WARNING oslo_messaging._drivers.amqpdriver [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 1007.343241] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-9eb14965-1725-434e-ae0a-8c0e6a976776" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.343531] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-9eb14965-1725-434e-ae0a-8c0e6a976776" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.343909] env[61995]: DEBUG nova.objects.instance [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'flavor' on Instance uuid 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.357957] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795079, 'name': CreateVM_Task, 'duration_secs': 0.598055} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.358582] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1007.359936] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.360124] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.360505] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1007.360960] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a0f7223-392c-48f8-9828-7f956a0b5a86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.365336] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1007.365336] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525652ae-b99f-cb75-53eb-78ee7e87a56c" [ 1007.365336] env[61995]: _type = "Task" [ 1007.365336] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.373465] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525652ae-b99f-cb75-53eb-78ee7e87a56c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.410032] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6a48a6ab-c7c1-4b0a-a9e6-825d1c426c4e tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.325s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.515576] env[61995]: DEBUG nova.objects.instance [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'flavor' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.623152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.775785] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.787308] env[61995]: DEBUG nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1007.816605] env[61995]: DEBUG nova.network.neutron [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updated VIF entry in instance network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1007.817131] env[61995]: DEBUG nova.network.neutron [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.843849] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f3d6f8-88aa-4943-ab54-9ac0c32840f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.855700] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88eab05e-0124-493b-91ee-af10d136def4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.890447] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4011012-b24c-4b4b-9951-53c8da93f57f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.898543] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525652ae-b99f-cb75-53eb-78ee7e87a56c, 'name': SearchDatastore_Task, 'duration_secs': 0.0109} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.900675] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.900973] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1007.901281] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.901507] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.901761] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1007.902092] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-263ae5d8-c19c-462d-b69b-b35fea317a95 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.904974] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6eab72-6b38-48b8-8f4b-182c03ce3739 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.921581] env[61995]: DEBUG nova.compute.provider_tree [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.926090] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1007.926090] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1007.926090] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acbd0dee-e9b3-48f5-befa-8853e746cb32 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.931487] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1007.931487] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5213bf18-4bca-3994-8bfe-17777d9dde6a" [ 1007.931487] env[61995]: _type = "Task" [ 1007.931487] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.940991] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5213bf18-4bca-3994-8bfe-17777d9dde6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.967404] env[61995]: DEBUG nova.objects.instance [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'pci_requests' on Instance uuid 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.022431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.023064] env[61995]: DEBUG oslo_concurrency.lockutils [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.023064] env[61995]: DEBUG nova.network.neutron [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1008.023064] env[61995]: DEBUG nova.objects.instance [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'info_cache' on Instance uuid 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.049597] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.049944] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.050200] env[61995]: DEBUG nova.compute.manager [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.050957] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de9bf02-227e-4c8c-8391-c6c814548359 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.060729] env[61995]: DEBUG nova.compute.manager [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1008.061488] env[61995]: DEBUG nova.objects.instance [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lazy-loading 'flavor' on Instance uuid 9e30eadd-5694-4fc7-8b54-2cf1d1571504 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.313067] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.319307] env[61995]: DEBUG oslo_concurrency.lockutils [req-616bda57-fd78-4c0c-b22b-2429d955b17b req-1e2a2a9d-41bb-4ff4-b2f4-92d80f94013a service nova] Releasing lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.426393] env[61995]: DEBUG nova.scheduler.client.report [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.446620] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5213bf18-4bca-3994-8bfe-17777d9dde6a, 'name': SearchDatastore_Task, 'duration_secs': 0.008215} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.447726] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f0e3c29-8075-45ba-be2e-b70db245606d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.455505] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1008.455505] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52024146-8581-8a82-2d06-5faa842c5c4d" [ 1008.455505] env[61995]: _type = "Task" [ 1008.455505] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.468488] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52024146-8581-8a82-2d06-5faa842c5c4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.470175] env[61995]: DEBUG nova.objects.base [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Object Instance<453f46c7-5bad-4ca2-b228-f76e62fbd03e> lazy-loaded attributes: flavor,pci_requests {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1008.470421] env[61995]: DEBUG nova.network.neutron [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1008.526680] env[61995]: DEBUG nova.objects.base [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Object Instance<229d966c-b70d-4a51-a176-2e88488d5c4b> lazy-loaded attributes: flavor,info_cache {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1008.559185] env[61995]: DEBUG nova.policy [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1008.567322] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1008.567609] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a6a2e42-78c5-4c33-a827-64bae17bb191 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.575959] env[61995]: DEBUG oslo_vmware.api [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1008.575959] env[61995]: value = "task-795080" [ 1008.575959] env[61995]: _type = "Task" [ 1008.575959] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.585034] env[61995]: DEBUG oslo_vmware.api [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795080, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.932124] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.841s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.934641] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.514s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.936228] env[61995]: INFO nova.compute.claims [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.953404] env[61995]: INFO nova.scheduler.client.report [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted allocations for instance 349667a3-6311-4c9d-bad2-ef4a5d1c52c8 [ 1008.968348] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52024146-8581-8a82-2d06-5faa842c5c4d, 'name': SearchDatastore_Task, 'duration_secs': 0.01515} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.968654] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.968929] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/66c56f7b-1fc7-4aed-8afc-350817e1ca48.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1008.969257] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-debff29d-79b6-432d-be9f-25eef319d44d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.976398] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1008.976398] env[61995]: value = "task-795081" [ 1008.976398] env[61995]: _type = "Task" [ 1008.976398] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.985189] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.085483] env[61995]: DEBUG oslo_vmware.api [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795080, 'name': PowerOffVM_Task, 'duration_secs': 0.20602} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.085782] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1009.085990] env[61995]: DEBUG nova.compute.manager [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.086761] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a13d89-adca-4990-a277-f40d248d7e5e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.303108] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Didn't find any instances for network info cache update. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1009.303400] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.303566] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.303714] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.304275] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.304275] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.304275] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.304275] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1009.304615] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.339433] env[61995]: DEBUG nova.network.neutron [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [{"id": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "address": "fa:16:3e:52:68:4b", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfc0bea1-d5", "ovs_interfaceid": "cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.466495] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d23c05e9-847d-4b59-aefa-b81d0a9dc9b3 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "349667a3-6311-4c9d-bad2-ef4a5d1c52c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.031s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.489593] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795081, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.598887] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cf2da119-869d-413a-b915-f6b55a0df0d6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.808269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.843227] env[61995]: DEBUG oslo_concurrency.lockutils [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-229d966c-b70d-4a51-a176-2e88488d5c4b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.988746] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611587} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.991686] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/66c56f7b-1fc7-4aed-8afc-350817e1ca48.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1009.992160] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1009.992819] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a3c2bbd-8aab-41b7-83e3-476f76e6bfa3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.005274] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1010.005274] env[61995]: value = "task-795082" [ 1010.005274] env[61995]: _type = "Task" [ 1010.005274] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.016088] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795082, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.194384] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b9370c-4d39-4614-a006-9a6d58d986c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.202345] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bba56fd-5d2d-48a7-a41c-396250c7ce7e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.232724] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3aeb32-9c49-4331-bcb3-8bd6f9073c3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.242381] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c4c5eb-ed7b-4c3e-bad9-732119dda3cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.255650] env[61995]: DEBUG nova.compute.provider_tree [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.296993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "8ae65d63-16c7-4c67-a8c2-73968f63457d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.296993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.296993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "8ae65d63-16c7-4c67-a8c2-73968f63457d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.296993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.296993] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.297736] env[61995]: INFO nova.compute.manager [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Terminating instance [ 1010.299741] env[61995]: DEBUG nova.compute.manager [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.300096] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.301063] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827e8834-363e-4243-8356-eb7730162d4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.308380] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.308753] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-382fa5c8-82e5-40f2-962f-263e59a9dad4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.318018] env[61995]: DEBUG oslo_vmware.api [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1010.318018] env[61995]: value = "task-795083" [ 1010.318018] env[61995]: _type = "Task" [ 1010.318018] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.324375] env[61995]: DEBUG oslo_vmware.api [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.347169] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1010.347479] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f827cd3b-0ee8-4008-b21e-eaf7fc776547 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.354617] env[61995]: DEBUG oslo_vmware.api [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1010.354617] env[61995]: value = "task-795084" [ 1010.354617] env[61995]: _type = "Task" [ 1010.354617] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.365053] env[61995]: DEBUG oslo_vmware.api [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795084, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.443857] env[61995]: DEBUG nova.network.neutron [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Successfully updated port: 9eb14965-1725-434e-ae0a-8c0e6a976776 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1010.518415] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795082, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083681} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.518805] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1010.519511] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5a8a04-4bf5-4a2d-b51f-575b23962cb8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.542385] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/66c56f7b-1fc7-4aed-8afc-350817e1ca48.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.542717] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0beed4d1-9ce5-4c4c-a76a-a21c618b320d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.562541] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1010.562541] env[61995]: value = "task-795085" [ 1010.562541] env[61995]: _type = "Task" [ 1010.562541] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.570874] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795085, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.762371] env[61995]: DEBUG nova.scheduler.client.report [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.826489] env[61995]: DEBUG oslo_vmware.api [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795083, 'name': PowerOffVM_Task, 'duration_secs': 0.232578} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.826794] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1010.827189] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1010.827352] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5ebef66-0d90-422f-a35f-9e530afb2711 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.864652] env[61995]: DEBUG oslo_vmware.api [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795084, 'name': PowerOnVM_Task, 'duration_secs': 0.419044} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.864950] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1010.865164] env[61995]: DEBUG nova.compute.manager [None req-61e2c491-9d01-4e5b-ab35-b87c3db2418e tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.865907] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf4a013-e52e-4174-b73d-d2cc263ea4bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.909906] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1010.910155] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1010.910362] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleting the datastore file [datastore2] 8ae65d63-16c7-4c67-a8c2-73968f63457d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.910945] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ede65e21-dae2-4a1e-b357-92e549e120dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.917501] env[61995]: DEBUG oslo_vmware.api [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1010.917501] env[61995]: value = "task-795087" [ 1010.917501] env[61995]: _type = "Task" [ 1010.917501] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.924822] env[61995]: DEBUG oslo_vmware.api [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795087, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.946818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.946996] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.947197] env[61995]: DEBUG nova.network.neutron [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.076421] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.086914] env[61995]: DEBUG nova.compute.manager [req-1b52475d-606a-43db-9f00-df624133ef9b req-ecc0ade2-cde4-4bc1-89e2-b6c3b912d1a5 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-vif-plugged-9eb14965-1725-434e-ae0a-8c0e6a976776 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.087336] env[61995]: DEBUG oslo_concurrency.lockutils [req-1b52475d-606a-43db-9f00-df624133ef9b req-ecc0ade2-cde4-4bc1-89e2-b6c3b912d1a5 service nova] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.087388] env[61995]: DEBUG oslo_concurrency.lockutils [req-1b52475d-606a-43db-9f00-df624133ef9b req-ecc0ade2-cde4-4bc1-89e2-b6c3b912d1a5 service nova] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.087548] env[61995]: DEBUG oslo_concurrency.lockutils [req-1b52475d-606a-43db-9f00-df624133ef9b req-ecc0ade2-cde4-4bc1-89e2-b6c3b912d1a5 service nova] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.087703] env[61995]: DEBUG nova.compute.manager [req-1b52475d-606a-43db-9f00-df624133ef9b req-ecc0ade2-cde4-4bc1-89e2-b6c3b912d1a5 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] No waiting events found dispatching network-vif-plugged-9eb14965-1725-434e-ae0a-8c0e6a976776 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.087875] env[61995]: WARNING nova.compute.manager [req-1b52475d-606a-43db-9f00-df624133ef9b req-ecc0ade2-cde4-4bc1-89e2-b6c3b912d1a5 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received unexpected event network-vif-plugged-9eb14965-1725-434e-ae0a-8c0e6a976776 for instance with vm_state active and task_state None. [ 1011.267471] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.268114] env[61995]: DEBUG nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.271071] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.230s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.274045] env[61995]: INFO nova.compute.claims [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.427332] env[61995]: DEBUG oslo_vmware.api [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795087, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129496} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.427592] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.427779] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1011.427957] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1011.428186] env[61995]: INFO nova.compute.manager [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1011.428436] env[61995]: DEBUG oslo.service.loopingcall [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.428623] env[61995]: DEBUG nova.compute.manager [-] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.428753] env[61995]: DEBUG nova.network.neutron [-] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1011.486512] env[61995]: WARNING nova.network.neutron [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 1011.487022] env[61995]: WARNING nova.network.neutron [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 1011.578416] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795085, 'name': ReconfigVM_Task, 'duration_secs': 0.773104} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.578903] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/66c56f7b-1fc7-4aed-8afc-350817e1ca48.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1011.579516] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43c7f669-e216-4bee-b28c-8b8553dc758b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.586560] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1011.586560] env[61995]: value = "task-795088" [ 1011.586560] env[61995]: _type = "Task" [ 1011.586560] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.596427] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795088, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.776960] env[61995]: DEBUG nova.compute.utils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1011.780321] env[61995]: DEBUG nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1011.780530] env[61995]: DEBUG nova.network.neutron [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1011.827299] env[61995]: DEBUG nova.policy [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90cf38026db34ee582d88a62ba1e0ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a383c3c42a2f4526ad1fba58b3e00807', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1011.979268] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "229d966c-b70d-4a51-a176-2e88488d5c4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.980130] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.980130] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.980130] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.980375] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.983554] env[61995]: DEBUG nova.network.neutron [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "address": "fa:16:3e:c5:f2:9f", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c9dcf57-62", "ovs_interfaceid": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9eb14965-1725-434e-ae0a-8c0e6a976776", "address": "fa:16:3e:2d:e4:19", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb14965-17", "ovs_interfaceid": "9eb14965-1725-434e-ae0a-8c0e6a976776", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.985141] env[61995]: INFO nova.compute.manager [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Terminating instance [ 1011.987083] env[61995]: DEBUG nova.compute.manager [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1011.987333] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1011.988271] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f2b85e-8068-4eb6-b539-cbf5edd7214f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.999191] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1011.999488] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb05c425-4859-4f31-b4a6-7402b515120a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.006888] env[61995]: DEBUG oslo_vmware.api [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1012.006888] env[61995]: value = "task-795089" [ 1012.006888] env[61995]: _type = "Task" [ 1012.006888] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.014697] env[61995]: DEBUG oslo_vmware.api [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.081158] env[61995]: DEBUG nova.compute.manager [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.086279] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33612184-44c0-488f-b7b5-2aad059e528f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.112572] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795088, 'name': Rename_Task, 'duration_secs': 0.315629} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.112956] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1012.113237] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25e65797-39f6-4d64-9236-7b7e3b468caf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.116275] env[61995]: DEBUG nova.network.neutron [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Successfully created port: 14a93316-278c-4140-a1e5-567a797d3a44 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.123021] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1012.123021] env[61995]: value = "task-795090" [ 1012.123021] env[61995]: _type = "Task" [ 1012.123021] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.131478] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.197443] env[61995]: DEBUG nova.network.neutron [-] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.281280] env[61995]: DEBUG nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.488225] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.488908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.489107] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.489928] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65f9dc9-0aab-4909-9584-39c3aca325c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.508697] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.508946] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.509146] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.509348] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.509502] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.509652] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.509862] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.510046] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.510226] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.510397] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.510581] env[61995]: DEBUG nova.virt.hardware [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.516681] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfiguring VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1012.522170] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ece3693c-a8a7-4db7-a349-0d85d99014f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.538547] env[61995]: DEBUG oslo_vmware.api [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795089, 'name': PowerOffVM_Task, 'duration_secs': 0.303368} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.539891] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1012.539891] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1012.540207] env[61995]: DEBUG oslo_vmware.api [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1012.540207] env[61995]: value = "task-795091" [ 1012.540207] env[61995]: _type = "Task" [ 1012.540207] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.540387] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f60b23c1-aea0-4166-b82a-7c646f081a40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.545602] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26e5ffb-d73d-4ca3-999b-962294d9618c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.552456] env[61995]: DEBUG oslo_vmware.api [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795091, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.554852] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9771ec-1164-4a5c-bf78-50878ace2ca1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.588964] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08120f1-18f6-451a-b2a8-0f47f1e0449d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.600017] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef09b8e-fce0-40b8-b8b5-300f0957f3cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.612132] env[61995]: DEBUG nova.compute.provider_tree [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.616847] env[61995]: INFO nova.compute.manager [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] instance snapshotting [ 1012.616988] env[61995]: WARNING nova.compute.manager [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 1012.619839] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0145e1-d9ba-4372-8d45-21e6b7f65d11 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.641560] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f875d9cc-2fc1-4b0e-92d8-5cd49ccaa453 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.647474] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795090, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.666756] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1012.666926] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1012.669411] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleting the datastore file [datastore2] 229d966c-b70d-4a51-a176-2e88488d5c4b {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.669411] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b81a594-6d5e-417e-ae71-0f39b46390d8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.675456] env[61995]: DEBUG oslo_vmware.api [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1012.675456] env[61995]: value = "task-795093" [ 1012.675456] env[61995]: _type = "Task" [ 1012.675456] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.682479] env[61995]: DEBUG oslo_vmware.api [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.700531] env[61995]: INFO nova.compute.manager [-] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Took 1.27 seconds to deallocate network for instance. [ 1013.051923] env[61995]: DEBUG oslo_vmware.api [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795091, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.116097] env[61995]: DEBUG nova.scheduler.client.report [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.126375] env[61995]: DEBUG nova.compute.manager [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-changed-9eb14965-1725-434e-ae0a-8c0e6a976776 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1013.126646] env[61995]: DEBUG nova.compute.manager [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing instance network info cache due to event network-changed-9eb14965-1725-434e-ae0a-8c0e6a976776. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1013.127009] env[61995]: DEBUG oslo_concurrency.lockutils [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.127274] env[61995]: DEBUG oslo_concurrency.lockutils [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.127536] env[61995]: DEBUG nova.network.neutron [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Refreshing network info cache for port 9eb14965-1725-434e-ae0a-8c0e6a976776 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1013.147382] env[61995]: DEBUG oslo_vmware.api [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795090, 'name': PowerOnVM_Task, 'duration_secs': 0.918745} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.147816] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1013.148936] env[61995]: INFO nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Took 9.41 seconds to spawn the instance on the hypervisor. [ 1013.148936] env[61995]: DEBUG nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.150502] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499ab6bc-66ef-4576-a241-ac8882fca8c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.156298] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1013.157032] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-74c507b6-c5fc-425a-b8f4-67eee870717a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.167283] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1013.167283] env[61995]: value = "task-795094" [ 1013.167283] env[61995]: _type = "Task" [ 1013.167283] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.175658] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795094, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.183483] env[61995]: DEBUG oslo_vmware.api [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215945} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.183718] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.183910] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1013.184123] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1013.184309] env[61995]: INFO nova.compute.manager [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1013.184557] env[61995]: DEBUG oslo.service.loopingcall [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.184749] env[61995]: DEBUG nova.compute.manager [-] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1013.184845] env[61995]: DEBUG nova.network.neutron [-] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1013.207317] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.292805] env[61995]: DEBUG nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.315322] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.315580] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.315733] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.315919] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.316250] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.316430] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.316651] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.316823] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.317805] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.317805] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.317805] env[61995]: DEBUG nova.virt.hardware [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.318255] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354e51d7-5107-4749-9256-7fb9ff3d8433 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.326525] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef4fbce-75e3-4d24-a94a-4e4064a6c78f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.552814] env[61995]: DEBUG oslo_vmware.api [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795091, 'name': ReconfigVM_Task, 'duration_secs': 0.622365} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.553290] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.553510] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfigured VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1013.614179] env[61995]: DEBUG nova.compute.manager [req-cd7bf954-e1bd-4fe5-aac2-e7d4164ae4af req-a396a1ed-2731-4e69-abbf-389aa67b43cd service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Received event network-vif-plugged-14a93316-278c-4140-a1e5-567a797d3a44 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1013.614179] env[61995]: DEBUG oslo_concurrency.lockutils [req-cd7bf954-e1bd-4fe5-aac2-e7d4164ae4af req-a396a1ed-2731-4e69-abbf-389aa67b43cd service nova] Acquiring lock "06886222-5f7f-482e-b5ee-afd7326f2c70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.614179] env[61995]: DEBUG oslo_concurrency.lockutils [req-cd7bf954-e1bd-4fe5-aac2-e7d4164ae4af req-a396a1ed-2731-4e69-abbf-389aa67b43cd service nova] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.614476] env[61995]: DEBUG oslo_concurrency.lockutils [req-cd7bf954-e1bd-4fe5-aac2-e7d4164ae4af req-a396a1ed-2731-4e69-abbf-389aa67b43cd service nova] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.614476] env[61995]: DEBUG nova.compute.manager [req-cd7bf954-e1bd-4fe5-aac2-e7d4164ae4af req-a396a1ed-2731-4e69-abbf-389aa67b43cd service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] No waiting events found dispatching network-vif-plugged-14a93316-278c-4140-a1e5-567a797d3a44 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1013.614548] env[61995]: WARNING nova.compute.manager [req-cd7bf954-e1bd-4fe5-aac2-e7d4164ae4af req-a396a1ed-2731-4e69-abbf-389aa67b43cd service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Received unexpected event network-vif-plugged-14a93316-278c-4140-a1e5-567a797d3a44 for instance with vm_state building and task_state spawning. [ 1013.622812] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.623304] env[61995]: DEBUG nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.626440] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.003s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.626440] env[61995]: DEBUG nova.objects.instance [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'resources' on Instance uuid 9627f278-73d6-45d3-b60d-776e373eef73 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.689999] env[61995]: INFO nova.compute.manager [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Took 15.20 seconds to build instance. [ 1013.693727] env[61995]: DEBUG nova.compute.manager [req-a9f3b089-9085-43fd-95f0-f6f928ce65e5 req-a6d467fa-534d-4ba2-ba5d-cdd6ec67acb0 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Received event network-vif-deleted-cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1013.693934] env[61995]: INFO nova.compute.manager [req-a9f3b089-9085-43fd-95f0-f6f928ce65e5 req-a6d467fa-534d-4ba2-ba5d-cdd6ec67acb0 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Neutron deleted interface cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01; detaching it from the instance and deleting it from the info cache [ 1013.694136] env[61995]: DEBUG nova.network.neutron [req-a9f3b089-9085-43fd-95f0-f6f928ce65e5 req-a6d467fa-534d-4ba2-ba5d-cdd6ec67acb0 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.699200] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795094, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.868017] env[61995]: DEBUG nova.network.neutron [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updated VIF entry in instance network info cache for port 9eb14965-1725-434e-ae0a-8c0e6a976776. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.868592] env[61995]: DEBUG nova.network.neutron [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "address": "fa:16:3e:c5:f2:9f", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c9dcf57-62", "ovs_interfaceid": "8c9dcf57-6289-4787-94d0-0d19f3a76395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9eb14965-1725-434e-ae0a-8c0e6a976776", "address": "fa:16:3e:2d:e4:19", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb14965-17", "ovs_interfaceid": "9eb14965-1725-434e-ae0a-8c0e6a976776", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.059955] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e93ccb1b-dcb1-4e71-8151-cd82a7dedfae tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-9eb14965-1725-434e-ae0a-8c0e6a976776" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.716s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.129871] env[61995]: DEBUG nova.compute.utils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.134913] env[61995]: DEBUG nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.134913] env[61995]: DEBUG nova.network.neutron [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1014.181718] env[61995]: DEBUG nova.network.neutron [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Successfully updated port: 14a93316-278c-4140-a1e5-567a797d3a44 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1014.186064] env[61995]: DEBUG nova.network.neutron [-] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.191048] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795094, 'name': CreateSnapshot_Task, 'duration_secs': 0.592826} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.191329] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1014.192096] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbfd62d-311c-49be-9a11-d4d52c535783 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.196287] env[61995]: DEBUG nova.policy [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90cf38026db34ee582d88a62ba1e0ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a383c3c42a2f4526ad1fba58b3e00807', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1014.198455] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fa598efd-9d8e-4ad9-859b-b0a1c35c39fe tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.718s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.206431] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59b8cd4e-82f4-4067-b499-cc1121c45bc3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.220062] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0d91d0-3ba1-4b67-b35e-a889d904c6e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.252028] env[61995]: DEBUG nova.compute.manager [req-a9f3b089-9085-43fd-95f0-f6f928ce65e5 req-a6d467fa-534d-4ba2-ba5d-cdd6ec67acb0 service nova] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Detach interface failed, port_id=cfc0bea1-d5e6-42e3-9d7c-f1aa4cf68f01, reason: Instance 229d966c-b70d-4a51-a176-2e88488d5c4b could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1014.371277] env[61995]: DEBUG oslo_concurrency.lockutils [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.371539] env[61995]: DEBUG nova.compute.manager [req-63639469-90ba-4c85-a7c2-691af32bf5b0 req-0c9a16a2-5c70-4ed5-a6df-2977f4f7af31 service nova] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Received event network-vif-deleted-0e1358a1-ed1a-453b-882f-a94d4440e000 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.395551] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a048017a-11f7-42d4-b0a6-09e9203dbe5d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.404022] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba4d57f-6d76-4d1d-8fc9-0ed1150fff1e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.436648] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1672ca-e859-4fae-bde5-9387cb035410 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.445286] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f35d329-905f-4bef-98cb-26c3bc47bb69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.458447] env[61995]: DEBUG nova.compute.provider_tree [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.468171] env[61995]: DEBUG nova.network.neutron [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Successfully created port: f5f8ea24-7208-4f4c-bec1-77e13198d09c {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.635280] env[61995]: DEBUG nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.690108] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.690108] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.690108] env[61995]: DEBUG nova.network.neutron [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.690312] env[61995]: INFO nova.compute.manager [-] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Took 1.51 seconds to deallocate network for instance. [ 1014.718123] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1014.718829] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b2630d4a-8d08-4918-97b7-db939bf26239 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.728396] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1014.728396] env[61995]: value = "task-795095" [ 1014.728396] env[61995]: _type = "Task" [ 1014.728396] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.740786] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795095, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.963152] env[61995]: DEBUG nova.scheduler.client.report [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.196833] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.206218] env[61995]: DEBUG nova.compute.manager [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Received event network-changed-14a93316-278c-4140-a1e5-567a797d3a44 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.206218] env[61995]: DEBUG nova.compute.manager [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Refreshing instance network info cache due to event network-changed-14a93316-278c-4140-a1e5-567a797d3a44. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1015.206218] env[61995]: DEBUG oslo_concurrency.lockutils [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] Acquiring lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.234084] env[61995]: DEBUG nova.network.neutron [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1015.242527] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795095, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.386780] env[61995]: DEBUG nova.network.neutron [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Updating instance_info_cache with network_info: [{"id": "14a93316-278c-4140-a1e5-567a797d3a44", "address": "fa:16:3e:73:e5:57", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14a93316-27", "ovs_interfaceid": "14a93316-278c-4140-a1e5-567a797d3a44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.470371] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.472981] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.697s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.474519] env[61995]: INFO nova.compute.claims [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.493592] env[61995]: INFO nova.scheduler.client.report [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted allocations for instance 9627f278-73d6-45d3-b60d-776e373eef73 [ 1015.645435] env[61995]: DEBUG nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.673385] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.673730] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.673966] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.674238] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.674479] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.674660] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.674929] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.675157] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.675422] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.675596] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.675809] env[61995]: DEBUG nova.virt.hardware [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.677137] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15da5594-73c8-4fe5-85b9-68db83fff4d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.687511] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cdd654-1894-4636-9265-71293a3ad218 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.710955] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-8c9dcf57-6289-4787-94d0-0d19f3a76395" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.711324] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-8c9dcf57-6289-4787-94d0-0d19f3a76395" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.740206] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795095, 'name': CloneVM_Task} progress is 95%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.889817] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.890173] env[61995]: DEBUG nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Instance network_info: |[{"id": "14a93316-278c-4140-a1e5-567a797d3a44", "address": "fa:16:3e:73:e5:57", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14a93316-27", "ovs_interfaceid": "14a93316-278c-4140-a1e5-567a797d3a44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.891526] env[61995]: DEBUG oslo_concurrency.lockutils [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] Acquired lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.891726] env[61995]: DEBUG nova.network.neutron [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Refreshing network info cache for port 14a93316-278c-4140-a1e5-567a797d3a44 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1015.892881] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:e5:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe20ef0e-0991-44d7-887d-08dddac0b56b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14a93316-278c-4140-a1e5-567a797d3a44', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.900351] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Creating folder: Project (a383c3c42a2f4526ad1fba58b3e00807). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1015.901495] env[61995]: DEBUG nova.compute.manager [req-4c0dccf0-b0ca-445f-9103-9b406032e108 req-8bc9bbd2-f7b0-4ea2-9bff-a075b48dc264 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Received event network-vif-plugged-f5f8ea24-7208-4f4c-bec1-77e13198d09c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.901693] env[61995]: DEBUG oslo_concurrency.lockutils [req-4c0dccf0-b0ca-445f-9103-9b406032e108 req-8bc9bbd2-f7b0-4ea2-9bff-a075b48dc264 service nova] Acquiring lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.901896] env[61995]: DEBUG oslo_concurrency.lockutils [req-4c0dccf0-b0ca-445f-9103-9b406032e108 req-8bc9bbd2-f7b0-4ea2-9bff-a075b48dc264 service nova] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.902077] env[61995]: DEBUG oslo_concurrency.lockutils [req-4c0dccf0-b0ca-445f-9103-9b406032e108 req-8bc9bbd2-f7b0-4ea2-9bff-a075b48dc264 service nova] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.902247] env[61995]: DEBUG nova.compute.manager [req-4c0dccf0-b0ca-445f-9103-9b406032e108 req-8bc9bbd2-f7b0-4ea2-9bff-a075b48dc264 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] No waiting events found dispatching network-vif-plugged-f5f8ea24-7208-4f4c-bec1-77e13198d09c {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.902414] env[61995]: WARNING nova.compute.manager [req-4c0dccf0-b0ca-445f-9103-9b406032e108 req-8bc9bbd2-f7b0-4ea2-9bff-a075b48dc264 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Received unexpected event network-vif-plugged-f5f8ea24-7208-4f4c-bec1-77e13198d09c for instance with vm_state building and task_state spawning. [ 1015.905649] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d8c6265-e802-4817-b40a-fba85363d6cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.918018] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Created folder: Project (a383c3c42a2f4526ad1fba58b3e00807) in parent group-v185203. [ 1015.918271] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Creating folder: Instances. Parent ref: group-v185393. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1015.918516] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3fef1595-ef2e-4d4f-92d3-573a793e0e33 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.929020] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Created folder: Instances in parent group-v185393. [ 1015.929343] env[61995]: DEBUG oslo.service.loopingcall [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.929598] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1015.929825] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ff901c3-0e92-406d-9b29-0354be8deeb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.951669] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.951669] env[61995]: value = "task-795098" [ 1015.951669] env[61995]: _type = "Task" [ 1015.951669] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.960332] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795098, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.986398] env[61995]: DEBUG nova.network.neutron [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Successfully updated port: f5f8ea24-7208-4f4c-bec1-77e13198d09c {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.002621] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55c47b0c-89e8-4de6-a9f7-dfb9b942b21d tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "9627f278-73d6-45d3-b60d-776e373eef73" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.459s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.146728] env[61995]: DEBUG nova.network.neutron [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Updated VIF entry in instance network info cache for port 14a93316-278c-4140-a1e5-567a797d3a44. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1016.147218] env[61995]: DEBUG nova.network.neutron [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Updating instance_info_cache with network_info: [{"id": "14a93316-278c-4140-a1e5-567a797d3a44", "address": "fa:16:3e:73:e5:57", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14a93316-27", "ovs_interfaceid": "14a93316-278c-4140-a1e5-567a797d3a44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.214214] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.214434] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.215344] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16340f37-18e4-4756-8252-ab0e6878f9df {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.237728] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328e93a4-17ec-4bcb-ac9a-2669b74fb2b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.245820] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795095, 'name': CloneVM_Task, 'duration_secs': 1.133959} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.263483] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Created linked-clone VM from snapshot [ 1016.269061] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfiguring VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1016.269804] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269cb961-d8dc-4fac-ba2a-c9b7fe06a6b6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.272719] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-401f4272-5fc9-44b1-b2cb-43af13f1d682 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.290803] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Uploading image 17255c28-b10b-4ad7-bafd-de8e81f4c675 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1016.294027] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1016.294027] env[61995]: value = "task-795099" [ 1016.294027] env[61995]: _type = "Task" [ 1016.294027] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.302074] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.316596] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1016.316596] env[61995]: value = "vm-185392" [ 1016.316596] env[61995]: _type = "VirtualMachine" [ 1016.316596] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1016.316866] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f5ac38fd-79d0-42d6-aca7-e031024acbc6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.323890] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease: (returnval){ [ 1016.323890] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b5274e-7a3e-1325-1888-87f50f38f819" [ 1016.323890] env[61995]: _type = "HttpNfcLease" [ 1016.323890] env[61995]: } obtained for exporting VM: (result){ [ 1016.323890] env[61995]: value = "vm-185392" [ 1016.323890] env[61995]: _type = "VirtualMachine" [ 1016.323890] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1016.324243] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the lease: (returnval){ [ 1016.324243] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b5274e-7a3e-1325-1888-87f50f38f819" [ 1016.324243] env[61995]: _type = "HttpNfcLease" [ 1016.324243] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1016.331050] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.331050] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b5274e-7a3e-1325-1888-87f50f38f819" [ 1016.331050] env[61995]: _type = "HttpNfcLease" [ 1016.331050] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1016.420357] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.420645] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.462141] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795098, 'name': CreateVM_Task, 'duration_secs': 0.37621} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.462308] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1016.463013] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.463196] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.463532] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1016.463777] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3be2c4f2-7a7b-4f4a-a9a2-6147238126fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.468088] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1016.468088] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526d192e-05ac-9ff5-196a-7b0f5ef530a1" [ 1016.468088] env[61995]: _type = "Task" [ 1016.468088] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.475401] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526d192e-05ac-9ff5-196a-7b0f5ef530a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.491076] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "refresh_cache-432f4d05-ea4d-402c-8934-3aa1aaf93d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.491222] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "refresh_cache-432f4d05-ea4d-402c-8934-3aa1aaf93d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.491364] env[61995]: DEBUG nova.network.neutron [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.649814] env[61995]: DEBUG oslo_concurrency.lockutils [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] Releasing lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.650109] env[61995]: DEBUG nova.compute.manager [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Received event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.650318] env[61995]: DEBUG nova.compute.manager [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing instance network info cache due to event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.650549] env[61995]: DEBUG oslo_concurrency.lockutils [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] Acquiring lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.650710] env[61995]: DEBUG oslo_concurrency.lockutils [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] Acquired lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.650889] env[61995]: DEBUG nova.network.neutron [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1016.662152] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ff1ddd-8ee6-473f-b89c-b7bba3713142 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.671813] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa806a59-97c3-46c4-a24e-63cb3117422d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.703336] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed129ed7-c90d-4a75-978f-38f4a660bc56 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.711828] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c111f76-b293-4446-8eb8-7bc15afa6916 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.726039] env[61995]: DEBUG nova.compute.provider_tree [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.804306] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.831614] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.831614] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b5274e-7a3e-1325-1888-87f50f38f819" [ 1016.831614] env[61995]: _type = "HttpNfcLease" [ 1016.831614] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1016.831925] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1016.831925] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b5274e-7a3e-1325-1888-87f50f38f819" [ 1016.831925] env[61995]: _type = "HttpNfcLease" [ 1016.831925] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1016.832672] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301bb8a5-d86e-489d-86bc-24ea8d2aac6e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.839893] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527254b4-a7b3-00ab-6152-18756fbf265a/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1016.840090] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527254b4-a7b3-00ab-6152-18756fbf265a/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1016.923680] env[61995]: DEBUG nova.compute.utils [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.929780] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2898ba28-6a38-4844-8cc4-97468d38cf15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.978242] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526d192e-05ac-9ff5-196a-7b0f5ef530a1, 'name': SearchDatastore_Task, 'duration_secs': 0.010358} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.978554] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.979139] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.979139] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.979273] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.979397] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.979653] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dabdec89-443c-4880-b4ef-8efa16f4340e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.987560] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.987743] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1016.988484] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-840e8567-aaa3-405d-9b96-448025a13c19 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.993072] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1016.993072] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52710c21-0b73-252e-c6d7-6106f76284cf" [ 1016.993072] env[61995]: _type = "Task" [ 1016.993072] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.004961] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52710c21-0b73-252e-c6d7-6106f76284cf, 'name': SearchDatastore_Task, 'duration_secs': 0.008904} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.006214] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b813e14-4d22-466e-b700-918e7199e48d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.010378] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1017.010378] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b38162-cc82-7091-5bef-18b61248a2a6" [ 1017.010378] env[61995]: _type = "Task" [ 1017.010378] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.017637] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b38162-cc82-7091-5bef-18b61248a2a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.027865] env[61995]: DEBUG nova.network.neutron [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.164236] env[61995]: DEBUG nova.network.neutron [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Updating instance_info_cache with network_info: [{"id": "f5f8ea24-7208-4f4c-bec1-77e13198d09c", "address": "fa:16:3e:b1:dd:21", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5f8ea24-72", "ovs_interfaceid": "f5f8ea24-7208-4f4c-bec1-77e13198d09c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.229018] env[61995]: DEBUG nova.scheduler.client.report [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.305446] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.427057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.458591] env[61995]: DEBUG nova.network.neutron [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updated VIF entry in instance network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.459369] env[61995]: DEBUG nova.network.neutron [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.521093] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b38162-cc82-7091-5bef-18b61248a2a6, 'name': SearchDatastore_Task, 'duration_secs': 0.008777} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.521367] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.521708] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 06886222-5f7f-482e-b5ee-afd7326f2c70/06886222-5f7f-482e-b5ee-afd7326f2c70.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1017.522123] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90995656-ceae-42bc-8fad-708e1f8d4cc3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.529863] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1017.529863] env[61995]: value = "task-795101" [ 1017.529863] env[61995]: _type = "Task" [ 1017.529863] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.538552] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795101, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.667313] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "refresh_cache-432f4d05-ea4d-402c-8934-3aa1aaf93d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.667754] env[61995]: DEBUG nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Instance network_info: |[{"id": "f5f8ea24-7208-4f4c-bec1-77e13198d09c", "address": "fa:16:3e:b1:dd:21", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5f8ea24-72", "ovs_interfaceid": "f5f8ea24-7208-4f4c-bec1-77e13198d09c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.668416] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:dd:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe20ef0e-0991-44d7-887d-08dddac0b56b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5f8ea24-7208-4f4c-bec1-77e13198d09c', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.678938] env[61995]: DEBUG oslo.service.loopingcall [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.679328] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1017.679565] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71b67d6f-861c-4575-bac4-c0bd7942536d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.699716] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.699716] env[61995]: value = "task-795102" [ 1017.699716] env[61995]: _type = "Task" [ 1017.699716] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.707877] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795102, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.736404] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.737014] env[61995]: DEBUG nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1017.739782] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.427s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.741258] env[61995]: INFO nova.compute.claims [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.806033] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.954643] env[61995]: DEBUG nova.compute.manager [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Received event network-changed-f5f8ea24-7208-4f4c-bec1-77e13198d09c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.955009] env[61995]: DEBUG nova.compute.manager [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Refreshing instance network info cache due to event network-changed-f5f8ea24-7208-4f4c-bec1-77e13198d09c. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1017.955310] env[61995]: DEBUG oslo_concurrency.lockutils [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] Acquiring lock "refresh_cache-432f4d05-ea4d-402c-8934-3aa1aaf93d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.955382] env[61995]: DEBUG oslo_concurrency.lockutils [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] Acquired lock "refresh_cache-432f4d05-ea4d-402c-8934-3aa1aaf93d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.955625] env[61995]: DEBUG nova.network.neutron [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Refreshing network info cache for port f5f8ea24-7208-4f4c-bec1-77e13198d09c {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.962354] env[61995]: DEBUG oslo_concurrency.lockutils [req-50be0d42-6fad-498f-8b8a-df7408c7cdf5 req-e02b353b-cb0c-4edd-8073-e65d335b002f service nova] Releasing lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.047032] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795101, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.209072] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795102, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.246235] env[61995]: DEBUG nova.compute.utils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.250094] env[61995]: DEBUG nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1018.250279] env[61995]: DEBUG nova.network.neutron [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1018.302256] env[61995]: DEBUG nova.policy [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90cf38026db34ee582d88a62ba1e0ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a383c3c42a2f4526ad1fba58b3e00807', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1018.309677] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.519684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.520116] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.520433] env[61995]: INFO nova.compute.manager [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Attaching volume 2453ed83-12b0-456e-9b54-9ba91f0671e0 to /dev/sdb [ 1018.529812] env[61995]: DEBUG nova.compute.manager [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1018.548827] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795101, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595447} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.553585] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 06886222-5f7f-482e-b5ee-afd7326f2c70/06886222-5f7f-482e-b5ee-afd7326f2c70.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1018.553889] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1018.554861] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8bc506b2-7a39-41b7-ae2c-a78374a468f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.568890] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1018.568890] env[61995]: value = "task-795103" [ 1018.568890] env[61995]: _type = "Task" [ 1018.568890] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.582302] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795103, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.596035] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d80434-81df-4aff-a1fc-b06f8791bc48 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.606165] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc48e3db-b658-4889-8c2a-d6b594af4485 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.620117] env[61995]: DEBUG nova.virt.block_device [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Updating existing volume attachment record: 044c1d3d-8e12-41a2-a5b6-45bfd4762f8e {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1018.629321] env[61995]: DEBUG nova.network.neutron [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Successfully created port: 97857bcc-fdda-481c-a35a-3c3c3ec6d995 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.711162] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795102, 'name': CreateVM_Task, 'duration_secs': 0.594478} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.711535] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1018.712065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.712252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.712618] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.712887] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47270977-30e0-4358-be9f-3cb3b173974a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.720134] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1018.720134] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52aeb979-7067-dc5f-b42f-42e154472c26" [ 1018.720134] env[61995]: _type = "Task" [ 1018.720134] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.730673] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52aeb979-7067-dc5f-b42f-42e154472c26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.732441] env[61995]: DEBUG nova.network.neutron [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Updated VIF entry in instance network info cache for port f5f8ea24-7208-4f4c-bec1-77e13198d09c. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.732795] env[61995]: DEBUG nova.network.neutron [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Updating instance_info_cache with network_info: [{"id": "f5f8ea24-7208-4f4c-bec1-77e13198d09c", "address": "fa:16:3e:b1:dd:21", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5f8ea24-72", "ovs_interfaceid": "f5f8ea24-7208-4f4c-bec1-77e13198d09c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.750947] env[61995]: DEBUG nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1018.810815] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.996833] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888826e1-6f25-4da6-9400-a78e410794f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.006215] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b309704a-5d6d-4466-82fb-6012c2bfa8c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.035634] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a1cb4c-8f8d-4b89-9816-7dbd592e9325 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.045781] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520712c0-67f3-41c7-87b6-8bc78efbd017 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.059299] env[61995]: DEBUG nova.compute.provider_tree [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.061024] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.077731] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795103, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090034} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.077996] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1019.078813] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41bb8bc-a061-4303-b36a-54a1946fe94a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.099832] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 06886222-5f7f-482e-b5ee-afd7326f2c70/06886222-5f7f-482e-b5ee-afd7326f2c70.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1019.100118] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23127438-ac32-4fa6-b544-55c92f4047e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.120665] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1019.120665] env[61995]: value = "task-795105" [ 1019.120665] env[61995]: _type = "Task" [ 1019.120665] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.129510] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.233843] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52aeb979-7067-dc5f-b42f-42e154472c26, 'name': SearchDatastore_Task, 'duration_secs': 0.015565} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.234208] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.234453] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.234703] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.234845] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.235031] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.235326] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-060821df-3444-4d33-b1a3-f08d9b198857 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.237802] env[61995]: DEBUG oslo_concurrency.lockutils [req-58200d1d-7e3c-4a32-972d-6502b981501e req-adae5750-a2bf-4451-a4aa-4ec58a715732 service nova] Releasing lock "refresh_cache-432f4d05-ea4d-402c-8934-3aa1aaf93d48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.244956] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.246021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1019.246021] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f59e53c0-916d-4cc2-9ff0-25893f40fc1d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.251263] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1019.251263] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52024b23-657c-7991-157c-28b4c1335fbb" [ 1019.251263] env[61995]: _type = "Task" [ 1019.251263] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.262532] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52024b23-657c-7991-157c-28b4c1335fbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.307750] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.563256] env[61995]: DEBUG nova.scheduler.client.report [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.637425] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795105, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.763031] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52024b23-657c-7991-157c-28b4c1335fbb, 'name': SearchDatastore_Task, 'duration_secs': 0.008879} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.764167] env[61995]: DEBUG nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1019.766073] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd35ac06-7699-4e68-8459-795a20641371 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.771901] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1019.771901] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526fafaa-6d40-b812-e424-5c570fc08c67" [ 1019.771901] env[61995]: _type = "Task" [ 1019.771901] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.780451] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526fafaa-6d40-b812-e424-5c570fc08c67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.788823] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1019.789096] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1019.789268] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.789459] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1019.789608] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.789759] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1019.789971] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1019.790150] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1019.790329] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1019.790560] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1019.790762] env[61995]: DEBUG nova.virt.hardware [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1019.791573] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb225b5-da38-4386-8fa1-b66645929235 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.798992] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f3f1f1-3451-4169-8420-3fb79e284c43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.818630] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.009410] env[61995]: DEBUG nova.compute.manager [req-7b7583a1-3ee2-42c4-8c48-b0361c8f4ba6 req-f2003aec-f595-45ca-97ff-29ccf9907c64 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Received event network-vif-plugged-97857bcc-fdda-481c-a35a-3c3c3ec6d995 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.009594] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b7583a1-3ee2-42c4-8c48-b0361c8f4ba6 req-f2003aec-f595-45ca-97ff-29ccf9907c64 service nova] Acquiring lock "a48d837d-1008-4248-bbe0-14321368e138-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.009809] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b7583a1-3ee2-42c4-8c48-b0361c8f4ba6 req-f2003aec-f595-45ca-97ff-29ccf9907c64 service nova] Lock "a48d837d-1008-4248-bbe0-14321368e138-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.009984] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b7583a1-3ee2-42c4-8c48-b0361c8f4ba6 req-f2003aec-f595-45ca-97ff-29ccf9907c64 service nova] Lock "a48d837d-1008-4248-bbe0-14321368e138-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.010398] env[61995]: DEBUG nova.compute.manager [req-7b7583a1-3ee2-42c4-8c48-b0361c8f4ba6 req-f2003aec-f595-45ca-97ff-29ccf9907c64 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] No waiting events found dispatching network-vif-plugged-97857bcc-fdda-481c-a35a-3c3c3ec6d995 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.010553] env[61995]: WARNING nova.compute.manager [req-7b7583a1-3ee2-42c4-8c48-b0361c8f4ba6 req-f2003aec-f595-45ca-97ff-29ccf9907c64 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Received unexpected event network-vif-plugged-97857bcc-fdda-481c-a35a-3c3c3ec6d995 for instance with vm_state building and task_state spawning. [ 1020.068806] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.069406] env[61995]: DEBUG nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1020.071985] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.264s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.072203] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.072361] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1020.072660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.866s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.072865] env[61995]: DEBUG nova.objects.instance [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'resources' on Instance uuid 8ae65d63-16c7-4c67-a8c2-73968f63457d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.078810] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b96f219-5282-41a3-81ba-a6f9245b9672 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.094537] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d01d67-8d38-45f5-bf9c-ccda2c16950b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.113428] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca96ab7-ea85-485c-b427-a92261626f1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.122318] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f897aaa-d82c-45d8-8615-350ee6eedbfb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.137301] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795105, 'name': ReconfigVM_Task, 'duration_secs': 0.531833} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.162120] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 06886222-5f7f-482e-b5ee-afd7326f2c70/06886222-5f7f-482e-b5ee-afd7326f2c70.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1020.163433] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180160MB free_disk=66GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1020.163573] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.163962] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b9ee21a-8cf5-4fb6-a0cc-2bd0cf02b4ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.171647] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1020.171647] env[61995]: value = "task-795106" [ 1020.171647] env[61995]: _type = "Task" [ 1020.171647] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.182190] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795106, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.286144] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526fafaa-6d40-b812-e424-5c570fc08c67, 'name': SearchDatastore_Task, 'duration_secs': 0.009984} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.286769] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.286769] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 432f4d05-ea4d-402c-8934-3aa1aaf93d48/432f4d05-ea4d-402c-8934-3aa1aaf93d48.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1020.287061] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de134ad9-4035-4f0c-8d63-2ac3ddd86f21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.293878] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1020.293878] env[61995]: value = "task-795107" [ 1020.293878] env[61995]: _type = "Task" [ 1020.293878] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.302420] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795107, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.314072] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.356343] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645370cc-87f3-4173-b48b-57c750152b15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.365182] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf1d9fd-888b-4b90-8e5a-81eb8d51417a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.398314] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f67d2b1-b8db-4324-b58d-68816db80cb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.406373] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef0fb6d-fe6f-4fa9-b470-f7e24411436d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.420182] env[61995]: DEBUG nova.compute.provider_tree [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.586283] env[61995]: DEBUG nova.compute.utils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.588604] env[61995]: DEBUG nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1020.588771] env[61995]: DEBUG nova.network.neutron [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.602656] env[61995]: DEBUG nova.network.neutron [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Successfully updated port: 97857bcc-fdda-481c-a35a-3c3c3ec6d995 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.634276] env[61995]: DEBUG nova.compute.manager [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Received event network-changed-97857bcc-fdda-481c-a35a-3c3c3ec6d995 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.634529] env[61995]: DEBUG nova.compute.manager [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Refreshing instance network info cache due to event network-changed-97857bcc-fdda-481c-a35a-3c3c3ec6d995. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1020.634788] env[61995]: DEBUG oslo_concurrency.lockutils [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] Acquiring lock "refresh_cache-a48d837d-1008-4248-bbe0-14321368e138" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.634946] env[61995]: DEBUG oslo_concurrency.lockutils [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] Acquired lock "refresh_cache-a48d837d-1008-4248-bbe0-14321368e138" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.635159] env[61995]: DEBUG nova.network.neutron [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Refreshing network info cache for port 97857bcc-fdda-481c-a35a-3c3c3ec6d995 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.637626] env[61995]: DEBUG nova.policy [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3efdbd0f8e0340dc8e4bfbf672d62f0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83064cbae17c429d8d084837635486da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1020.684035] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795106, 'name': Rename_Task, 'duration_secs': 0.217739} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.684275] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1020.684571] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e40e932-3540-4d4f-832f-efc5701277e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.694805] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1020.694805] env[61995]: value = "task-795108" [ 1020.694805] env[61995]: _type = "Task" [ 1020.694805] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.707293] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.806029] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795107, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435155} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.806029] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 432f4d05-ea4d-402c-8934-3aa1aaf93d48/432f4d05-ea4d-402c-8934-3aa1aaf93d48.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1020.806029] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.806029] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de2439b6-144a-4440-a418-1b2f26d2304a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.817023] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.817023] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1020.817023] env[61995]: value = "task-795109" [ 1020.817023] env[61995]: _type = "Task" [ 1020.817023] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.826965] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795109, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.923734] env[61995]: DEBUG nova.scheduler.client.report [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.969252] env[61995]: DEBUG nova.network.neutron [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Successfully created port: 505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1021.092235] env[61995]: DEBUG nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1021.104755] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "refresh_cache-a48d837d-1008-4248-bbe0-14321368e138" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.177384] env[61995]: DEBUG nova.network.neutron [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1021.205877] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795108, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.317949] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.318931] env[61995]: DEBUG nova.network.neutron [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.328472] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795109, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.144299} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.328768] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.329610] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b7e638-8d97-4664-b716-9a4fe927bd2a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.352497] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 432f4d05-ea4d-402c-8934-3aa1aaf93d48/432f4d05-ea4d-402c-8934-3aa1aaf93d48.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.353763] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-524ddf5b-73a9-4e0e-89b0-e4ff02e259e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.375123] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1021.375123] env[61995]: value = "task-795111" [ 1021.375123] env[61995]: _type = "Task" [ 1021.375123] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.384216] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795111, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.428390] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.355s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.431132] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.234s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.431354] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.433316] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.372s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.455501] env[61995]: INFO nova.scheduler.client.report [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted allocations for instance 8ae65d63-16c7-4c67-a8c2-73968f63457d [ 1021.460025] env[61995]: INFO nova.scheduler.client.report [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted allocations for instance 229d966c-b70d-4a51-a176-2e88488d5c4b [ 1021.705817] env[61995]: DEBUG oslo_vmware.api [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795108, 'name': PowerOnVM_Task, 'duration_secs': 0.664935} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.707070] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1021.707070] env[61995]: INFO nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Took 8.41 seconds to spawn the instance on the hypervisor. [ 1021.707070] env[61995]: DEBUG nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.707661] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61eda25-340c-423d-8f7c-db1b90a87144 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.818454] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.825314] env[61995]: DEBUG oslo_concurrency.lockutils [req-2bd794f7-45a9-429d-9867-b1cc7d5b9d75 req-a7cafdc6-13be-4c9f-9929-a73148e6d296 service nova] Releasing lock "refresh_cache-a48d837d-1008-4248-bbe0-14321368e138" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.825314] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "refresh_cache-a48d837d-1008-4248-bbe0-14321368e138" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.825314] env[61995]: DEBUG nova.network.neutron [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.885320] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795111, 'name': ReconfigVM_Task, 'duration_secs': 0.304821} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.885610] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 432f4d05-ea4d-402c-8934-3aa1aaf93d48/432f4d05-ea4d-402c-8934-3aa1aaf93d48.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.886405] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2994a60-c280-44ab-93a3-6f4cfb891dc9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.893013] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1021.893013] env[61995]: value = "task-795112" [ 1021.893013] env[61995]: _type = "Task" [ 1021.893013] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.902657] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795112, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.938576] env[61995]: INFO nova.compute.claims [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.965018] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5abd9308-48bd-41a7-b360-e697a0bd98e7 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "8ae65d63-16c7-4c67-a8c2-73968f63457d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.670s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.968763] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a9cfd1d5-ec45-4537-b047-1754e02df684 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "229d966c-b70d-4a51-a176-2e88488d5c4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.989s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.101509] env[61995]: DEBUG nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1022.144896] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1022.146114] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1022.146794] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.146881] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.147081] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.147275] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.147507] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.147701] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.147877] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.148057] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.148327] env[61995]: DEBUG nova.virt.hardware [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.150234] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cfa1301-cb7f-4486-a476-3e62af25324f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.160094] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ba8c73-204d-4179-bd38-faf9fdb70236 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.224770] env[61995]: INFO nova.compute.manager [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Took 15.83 seconds to build instance. [ 1022.317854] env[61995]: DEBUG oslo_vmware.api [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795099, 'name': ReconfigVM_Task, 'duration_secs': 5.897153} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.318160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.318386] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Reconfigured VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1022.378519] env[61995]: DEBUG nova.network.neutron [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1022.403392] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795112, 'name': Rename_Task, 'duration_secs': 0.147657} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.404038] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1022.404038] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1b5205a-1b83-43a9-aa5a-6d44b5ecbdf8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.410146] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1022.410146] env[61995]: value = "task-795113" [ 1022.410146] env[61995]: _type = "Task" [ 1022.410146] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.418306] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795113, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.445095] env[61995]: INFO nova.compute.resource_tracker [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating resource usage from migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2 [ 1022.654710] env[61995]: DEBUG nova.network.neutron [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Successfully updated port: 505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.727632] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ff8b5277-8da5-4786-9b7b-41744de6c957 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.343s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.772046] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a79db1-a94e-449b-9749-5ad03f390e82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.782327] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f81472a-c9f5-4c76-8cb7-7296b638f8d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.820372] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103a5e2e-1ab8-4db6-adba-e79db2035ac3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.828743] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8a4065-129c-4d72-b94f-31585795918f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.836335] env[61995]: DEBUG nova.compute.manager [req-0acfc747-1c7e-4f0a-b206-2e1bbc61358a req-f09da6c7-0d23-4f1a-994b-549770fc4079 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Received event network-vif-plugged-505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.836644] env[61995]: DEBUG oslo_concurrency.lockutils [req-0acfc747-1c7e-4f0a-b206-2e1bbc61358a req-f09da6c7-0d23-4f1a-994b-549770fc4079 service nova] Acquiring lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.836873] env[61995]: DEBUG oslo_concurrency.lockutils [req-0acfc747-1c7e-4f0a-b206-2e1bbc61358a req-f09da6c7-0d23-4f1a-994b-549770fc4079 service nova] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.837062] env[61995]: DEBUG oslo_concurrency.lockutils [req-0acfc747-1c7e-4f0a-b206-2e1bbc61358a req-f09da6c7-0d23-4f1a-994b-549770fc4079 service nova] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.837247] env[61995]: DEBUG nova.compute.manager [req-0acfc747-1c7e-4f0a-b206-2e1bbc61358a req-f09da6c7-0d23-4f1a-994b-549770fc4079 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] No waiting events found dispatching network-vif-plugged-505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.837439] env[61995]: WARNING nova.compute.manager [req-0acfc747-1c7e-4f0a-b206-2e1bbc61358a req-f09da6c7-0d23-4f1a-994b-549770fc4079 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Received unexpected event network-vif-plugged-505624a2-00f0-4314-b814-28c5bc7ac5d9 for instance with vm_state building and task_state spawning. [ 1022.848844] env[61995]: DEBUG nova.compute.provider_tree [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.923524] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795113, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.969046] env[61995]: DEBUG nova.network.neutron [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Updating instance_info_cache with network_info: [{"id": "97857bcc-fdda-481c-a35a-3c3c3ec6d995", "address": "fa:16:3e:78:e5:f4", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97857bcc-fd", "ovs_interfaceid": "97857bcc-fdda-481c-a35a-3c3c3ec6d995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.158276] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.158955] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.158955] env[61995]: DEBUG nova.network.neutron [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.186691] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1023.186966] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185397', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'name': 'volume-2453ed83-12b0-456e-9b54-9ba91f0671e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba89ee63-e293-47e1-90ab-7b8e72dd1b50', 'attached_at': '', 'detached_at': '', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'serial': '2453ed83-12b0-456e-9b54-9ba91f0671e0'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1023.187895] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8490b254-973e-4765-95b7-367916aa9c84 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.212962] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ace045-da3b-4f1b-a4e6-cb23ed5d9302 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.217706] env[61995]: DEBUG nova.compute.manager [req-f324547e-a401-42ca-8fd1-93f5a08d974b req-5e25cf7c-b9a2-4536-a2ea-85c53059e803 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-vif-deleted-8c9dcf57-6289-4787-94d0-0d19f3a76395 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1023.217706] env[61995]: INFO nova.compute.manager [req-f324547e-a401-42ca-8fd1-93f5a08d974b req-5e25cf7c-b9a2-4536-a2ea-85c53059e803 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Neutron deleted interface 8c9dcf57-6289-4787-94d0-0d19f3a76395; detaching it from the instance and deleting it from the info cache [ 1023.218615] env[61995]: DEBUG nova.network.neutron [req-f324547e-a401-42ca-8fd1-93f5a08d974b req-5e25cf7c-b9a2-4536-a2ea-85c53059e803 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9eb14965-1725-434e-ae0a-8c0e6a976776", "address": "fa:16:3e:2d:e4:19", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9eb14965-17", "ovs_interfaceid": "9eb14965-1725-434e-ae0a-8c0e6a976776", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.244073] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] volume-2453ed83-12b0-456e-9b54-9ba91f0671e0/volume-2453ed83-12b0-456e-9b54-9ba91f0671e0.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.245258] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-170c3b73-8a32-4f71-a8b0-c1773f7a17b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.267767] env[61995]: DEBUG oslo_vmware.api [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1023.267767] env[61995]: value = "task-795114" [ 1023.267767] env[61995]: _type = "Task" [ 1023.267767] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.277534] env[61995]: DEBUG oslo_vmware.api [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.305939] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "e8bde834-8f55-48d6-8623-1c6e43072c23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.306286] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.354985] env[61995]: DEBUG nova.scheduler.client.report [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.423579] env[61995]: DEBUG oslo_vmware.api [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795113, 'name': PowerOnVM_Task, 'duration_secs': 0.579431} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.423939] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1023.424240] env[61995]: INFO nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Took 7.78 seconds to spawn the instance on the hypervisor. [ 1023.424679] env[61995]: DEBUG nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.425617] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1af49e8-7171-4427-aff2-1ac04a58bdfe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.472754] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "refresh_cache-a48d837d-1008-4248-bbe0-14321368e138" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.473106] env[61995]: DEBUG nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Instance network_info: |[{"id": "97857bcc-fdda-481c-a35a-3c3c3ec6d995", "address": "fa:16:3e:78:e5:f4", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97857bcc-fd", "ovs_interfaceid": "97857bcc-fdda-481c-a35a-3c3c3ec6d995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1023.473559] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:e5:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe20ef0e-0991-44d7-887d-08dddac0b56b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97857bcc-fdda-481c-a35a-3c3c3ec6d995', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.481164] env[61995]: DEBUG oslo.service.loopingcall [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.482116] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a48d837d-1008-4248-bbe0-14321368e138] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1023.482359] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da8352e5-f288-44ab-bc82-7cad001ec06f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.503384] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1023.503384] env[61995]: value = "task-795115" [ 1023.503384] env[61995]: _type = "Task" [ 1023.503384] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.512898] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795115, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.697848] env[61995]: DEBUG nova.network.neutron [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.720797] env[61995]: DEBUG oslo_concurrency.lockutils [req-f324547e-a401-42ca-8fd1-93f5a08d974b req-5e25cf7c-b9a2-4536-a2ea-85c53059e803 service nova] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.720979] env[61995]: DEBUG oslo_concurrency.lockutils [req-f324547e-a401-42ca-8fd1-93f5a08d974b req-5e25cf7c-b9a2-4536-a2ea-85c53059e803 service nova] Acquired lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.722156] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5743a2a8-f175-4d08-84ea-bfd4db9bdeda {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.741514] env[61995]: DEBUG oslo_concurrency.lockutils [req-f324547e-a401-42ca-8fd1-93f5a08d974b req-5e25cf7c-b9a2-4536-a2ea-85c53059e803 service nova] Releasing lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.741809] env[61995]: WARNING nova.compute.manager [req-f324547e-a401-42ca-8fd1-93f5a08d974b req-5e25cf7c-b9a2-4536-a2ea-85c53059e803 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Detach interface failed, port_id=8c9dcf57-6289-4787-94d0-0d19f3a76395, reason: No device with interface-id 8c9dcf57-6289-4787-94d0-0d19f3a76395 exists on VM: nova.exception.NotFound: No device with interface-id 8c9dcf57-6289-4787-94d0-0d19f3a76395 exists on VM [ 1023.781857] env[61995]: DEBUG oslo_vmware.api [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.808712] env[61995]: DEBUG nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.861143] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.428s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.861438] env[61995]: INFO nova.compute.manager [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Migrating [ 1023.871681] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.708s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.884263] env[61995]: DEBUG nova.network.neutron [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Updating instance_info_cache with network_info: [{"id": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "address": "fa:16:3e:52:c6:6f", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap505624a2-00", "ovs_interfaceid": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.947599] env[61995]: INFO nova.compute.manager [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Took 16.92 seconds to build instance. [ 1023.994596] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "83369251-b00f-4595-bf98-28a3ec84f037" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.994941] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.018626] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795115, 'name': CreateVM_Task, 'duration_secs': 0.46374} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.018626] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a48d837d-1008-4248-bbe0-14321368e138] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.018626] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.018626] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.019201] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.019535] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86985823-85b3-431c-a56e-4d8f87443935 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.025176] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1024.025176] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5289bd7e-c49e-1e15-42ff-b2c918b9a591" [ 1024.025176] env[61995]: _type = "Task" [ 1024.025176] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.034634] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5289bd7e-c49e-1e15-42ff-b2c918b9a591, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.055136] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.055412] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.055723] env[61995]: DEBUG nova.network.neutron [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.279471] env[61995]: DEBUG oslo_vmware.api [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795114, 'name': ReconfigVM_Task, 'duration_secs': 0.646129} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.279839] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Reconfigured VM instance instance-00000052 to attach disk [datastore2] volume-2453ed83-12b0-456e-9b54-9ba91f0671e0/volume-2453ed83-12b0-456e-9b54-9ba91f0671e0.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.285496] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3f8f9fb-d8d3-489e-b091-17f3bfb66e02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.302729] env[61995]: DEBUG oslo_vmware.api [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1024.302729] env[61995]: value = "task-795116" [ 1024.302729] env[61995]: _type = "Task" [ 1024.302729] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.318691] env[61995]: DEBUG oslo_vmware.api [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795116, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.339701] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.395910] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.396212] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.396403] env[61995]: DEBUG nova.network.neutron [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.397964] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.397964] env[61995]: DEBUG nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Instance network_info: |[{"id": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "address": "fa:16:3e:52:c6:6f", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap505624a2-00", "ovs_interfaceid": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1024.399597] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:c6:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '505624a2-00f0-4314-b814-28c5bc7ac5d9', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.408254] env[61995]: DEBUG oslo.service.loopingcall [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.409333] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1024.409535] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0e4250f-ce8b-448a-b18f-5aeeaebd05bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.431047] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.431047] env[61995]: value = "task-795117" [ 1024.431047] env[61995]: _type = "Task" [ 1024.431047] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.441056] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795117, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.449889] env[61995]: DEBUG oslo_concurrency.lockutils [None req-14a6c8f6-73a7-4a52-a5f2-b9042ff5e4e2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.434s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.499084] env[61995]: DEBUG nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1024.539124] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5289bd7e-c49e-1e15-42ff-b2c918b9a591, 'name': SearchDatastore_Task, 'duration_secs': 0.024959} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.540236] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.540517] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.540794] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.540952] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.541176] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.541739] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5415ff68-2b22-4ea2-ad5e-4c28f4251f05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.550906] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.551158] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1024.551972] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbe899e2-4114-4276-9788-873a54f46c94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.558149] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1024.558149] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521d7cbe-d299-1249-ea9d-0a8701ce438e" [ 1024.558149] env[61995]: _type = "Task" [ 1024.558149] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.569120] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521d7cbe-d299-1249-ea9d-0a8701ce438e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.792382] env[61995]: INFO nova.network.neutron [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Port 9eb14965-1725-434e-ae0a-8c0e6a976776 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1024.792815] env[61995]: DEBUG nova.network.neutron [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [{"id": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "address": "fa:16:3e:ca:10:01", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdaef75a-14", "ovs_interfaceid": "cdaef75a-1442-4f8d-8eab-340da21a74d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.813474] env[61995]: DEBUG oslo_vmware.api [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795116, 'name': ReconfigVM_Task, 'duration_secs': 0.17279} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.813783] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185397', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'name': 'volume-2453ed83-12b0-456e-9b54-9ba91f0671e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba89ee63-e293-47e1-90ab-7b8e72dd1b50', 'attached_at': '', 'detached_at': '', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'serial': '2453ed83-12b0-456e-9b54-9ba91f0671e0'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1024.820179] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.820614] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.820883] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.821140] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.821380] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.823686] env[61995]: INFO nova.compute.manager [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Terminating instance [ 1024.825875] env[61995]: DEBUG nova.compute.manager [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.826135] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1024.827021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28acd350-eb61-4e5a-b6aa-a6522f89c5ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.838035] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1024.838035] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cb8946c-ad67-4f9b-a6a0-bc236d9cc49f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.845517] env[61995]: DEBUG oslo_vmware.api [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1024.845517] env[61995]: value = "task-795118" [ 1024.845517] env[61995]: _type = "Task" [ 1024.845517] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.855102] env[61995]: DEBUG oslo_vmware.api [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795118, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.858914] env[61995]: DEBUG nova.compute.manager [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Received event network-changed-505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1024.859102] env[61995]: DEBUG nova.compute.manager [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Refreshing instance network info cache due to event network-changed-505624a2-00f0-4314-b814-28c5bc7ac5d9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1024.859387] env[61995]: DEBUG oslo_concurrency.lockutils [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] Acquiring lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.859535] env[61995]: DEBUG oslo_concurrency.lockutils [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] Acquired lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.859706] env[61995]: DEBUG nova.network.neutron [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Refreshing network info cache for port 505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1024.900178] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Applying migration context for instance d9c53806-68c0-4872-a262-36bc05573674 as it has an incoming, in-progress migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2. Migration status is pre-migrating {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1024.901989] env[61995]: INFO nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating resource usage from migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2 [ 1024.925419] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b40b1866-ced3-40ef-9ab7-30d58b75f288 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.925597] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ed952baf-f1ef-4ec0-8a15-928287190456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.925704] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e835ac53-29af-4bd0-b186-5c6270ccf760 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.925821] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 453f46c7-5bad-4ca2-b228-f76e62fbd03e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.925937] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 24b974c0-9567-46a3-96f6-fac1e602f46e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.926110] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 5cb4c1d9-e661-48b4-8600-1b823d63ddbb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.926289] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ba89ee63-e293-47e1-90ab-7b8e72dd1b50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.926416] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 9e30eadd-5694-4fc7-8b54-2cf1d1571504 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.926544] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 66c56f7b-1fc7-4aed-8afc-350817e1ca48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.926656] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 06886222-5f7f-482e-b5ee-afd7326f2c70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.926770] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 432f4d05-ea4d-402c-8934-3aa1aaf93d48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.926895] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance a48d837d-1008-4248-bbe0-14321368e138 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.927021] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 410eac71-20cd-4a6d-9b78-e11fa72d74e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.927135] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1024.927258] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d9c53806-68c0-4872-a262-36bc05573674 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.942757] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795117, 'name': CreateVM_Task, 'duration_secs': 0.50614} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.943492] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.944207] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.944400] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.944729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.945224] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a2f5445-5c28-41db-8a1e-ac2298288d4f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.949956] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1024.949956] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526f1e3a-2faf-d18e-bd90-df9214615dba" [ 1024.949956] env[61995]: _type = "Task" [ 1024.949956] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.958565] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526f1e3a-2faf-d18e-bd90-df9214615dba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.018428] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.073514] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521d7cbe-d299-1249-ea9d-0a8701ce438e, 'name': SearchDatastore_Task, 'duration_secs': 0.015813} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.074052] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d02e2424-9975-4d8c-a448-30905611db14 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.085393] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1025.085393] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f201b6-0d06-c308-92b8-f10596167a30" [ 1025.085393] env[61995]: _type = "Task" [ 1025.085393] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.096506] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f201b6-0d06-c308-92b8-f10596167a30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.212147] env[61995]: DEBUG nova.network.neutron [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.242915] env[61995]: DEBUG nova.compute.manager [req-dd3de2fd-6646-4ba0-bb2f-7d6cf0b57df2 req-b9e6e8f7-a85c-415f-9c82-3490213bd967 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-vif-deleted-9eb14965-1725-434e-ae0a-8c0e6a976776 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.296740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-453f46c7-5bad-4ca2-b228-f76e62fbd03e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.357334] env[61995]: DEBUG oslo_vmware.api [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795118, 'name': PowerOffVM_Task, 'duration_secs': 0.28904} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.357749] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1025.358095] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1025.358953] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-364b0797-017c-4b19-8679-bb5fc6b7b348 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.433304] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e8bde834-8f55-48d6-8623-1c6e43072c23 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1025.462294] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526f1e3a-2faf-d18e-bd90-df9214615dba, 'name': SearchDatastore_Task, 'duration_secs': 0.023055} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.463029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.463029] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.463213] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.513017] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1025.513277] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1025.513491] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleting the datastore file [datastore1] 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.513751] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d22c35d9-f429-4ac1-93e1-fb42f902ce5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.525614] env[61995]: DEBUG oslo_vmware.api [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1025.525614] env[61995]: value = "task-795120" [ 1025.525614] env[61995]: _type = "Task" [ 1025.525614] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.535883] env[61995]: DEBUG oslo_vmware.api [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795120, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.601409] env[61995]: DEBUG nova.network.neutron [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Updated VIF entry in instance network info cache for port 505624a2-00f0-4314-b814-28c5bc7ac5d9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1025.601859] env[61995]: DEBUG nova.network.neutron [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Updating instance_info_cache with network_info: [{"id": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "address": "fa:16:3e:52:c6:6f", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap505624a2-00", "ovs_interfaceid": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.603198] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f201b6-0d06-c308-92b8-f10596167a30, 'name': SearchDatastore_Task, 'duration_secs': 0.01382} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.603399] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.603662] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] a48d837d-1008-4248-bbe0-14321368e138/a48d837d-1008-4248-bbe0-14321368e138.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1025.604216] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.604437] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.604672] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d7d9dfb-1041-4711-b618-be5492b3bd47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.606789] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7b2170e-585a-4550-a63c-6b40a9b97fe7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.618234] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527254b4-a7b3-00ab-6152-18756fbf265a/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1025.619233] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1339ef15-88af-4e25-8b69-6a4f0e75a9ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.623435] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1025.623435] env[61995]: value = "task-795121" [ 1025.623435] env[61995]: _type = "Task" [ 1025.623435] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.629576] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527254b4-a7b3-00ab-6152-18756fbf265a/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1025.629770] env[61995]: ERROR oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527254b4-a7b3-00ab-6152-18756fbf265a/disk-0.vmdk due to incomplete transfer. [ 1025.630019] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.630211] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1025.631283] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-62f40e4f-b555-4493-98f6-5f6a99e93f25 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.632801] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e639f51-769c-4c7d-a56c-ba69b019d91a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.638458] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795121, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.642630] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1025.642630] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52182cc3-19fd-b1d2-7145-3dd20372c57f" [ 1025.642630] env[61995]: _type = "Task" [ 1025.642630] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.647014] env[61995]: DEBUG oslo_vmware.rw_handles [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527254b4-a7b3-00ab-6152-18756fbf265a/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1025.647302] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Uploaded image 17255c28-b10b-4ad7-bafd-de8e81f4c675 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1025.649948] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1025.650543] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f4af9072-19f9-4f34-b232-67ae0599385c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.659623] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52182cc3-19fd-b1d2-7145-3dd20372c57f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.661193] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1025.661193] env[61995]: value = "task-795122" [ 1025.661193] env[61995]: _type = "Task" [ 1025.661193] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.676384] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795122, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.715053] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.800875] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fb61c61e-84e9-4eed-907c-5ffb2d063aa2 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-453f46c7-5bad-4ca2-b228-f76e62fbd03e-8c9dcf57-6289-4787-94d0-0d19f3a76395" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.089s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.853797] env[61995]: DEBUG nova.objects.instance [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'flavor' on Instance uuid ba89ee63-e293-47e1-90ab-7b8e72dd1b50 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.937057] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 83369251-b00f-4595-bf98-28a3ec84f037 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1025.937437] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1025.937578] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3520MB phys_disk=200GB used_disk=15GB total_vcpus=48 used_vcpus=15 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1026.039892] env[61995]: DEBUG oslo_vmware.api [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795120, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237961} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.040254] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.040531] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1026.040739] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1026.040924] env[61995]: INFO nova.compute.manager [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1026.041200] env[61995]: DEBUG oslo.service.loopingcall [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.041412] env[61995]: DEBUG nova.compute.manager [-] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.041510] env[61995]: DEBUG nova.network.neutron [-] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1026.105827] env[61995]: DEBUG oslo_concurrency.lockutils [req-3c4a04fc-2df6-47de-a42b-499799f78ce3 req-126386f6-d2bc-4d7b-bb4d-b0155d1d2bd5 service nova] Releasing lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.142890] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795121, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.144371] env[61995]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 9eb14965-1725-434e-ae0a-8c0e6a976776 could not be found.", "detail": ""}} {{(pid=61995) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1026.144371] env[61995]: DEBUG nova.network.neutron [-] Unable to show port 9eb14965-1725-434e-ae0a-8c0e6a976776 as it no longer exists. {{(pid=61995) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1026.158156] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52182cc3-19fd-b1d2-7145-3dd20372c57f, 'name': SearchDatastore_Task, 'duration_secs': 0.014865} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.159222] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db26947d-5bed-458f-9cfe-77512bfe2478 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.171435] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1026.171435] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5287e46a-2d75-69f8-c91c-4f3377476a1a" [ 1026.171435] env[61995]: _type = "Task" [ 1026.171435] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.180503] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795122, 'name': Destroy_Task, 'duration_secs': 0.513197} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.181580] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Destroyed the VM [ 1026.181580] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1026.181810] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6b7a78fd-028c-4c2f-9b4d-abddbc23bcd6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.186315] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5287e46a-2d75-69f8-c91c-4f3377476a1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.192187] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1026.192187] env[61995]: value = "task-795123" [ 1026.192187] env[61995]: _type = "Task" [ 1026.192187] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.204731] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795123, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.208728] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d12d19-84a0-4b1e-8480-77cb3c4ac884 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.217427] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c4a7e9-6b4b-4a0c-b9a6-90c2686f3e68 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.252027] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dcebf1-7824-4bbf-8cba-b090d4697462 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.261483] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f61b52e-1f8a-47e9-b51b-7c50700eff91 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.279278] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.359968] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8473da2c-7233-4ef1-8c5c-645ac6abb982 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.840s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.637092] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795121, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588738} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.637363] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] a48d837d-1008-4248-bbe0-14321368e138/a48d837d-1008-4248-bbe0-14321368e138.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1026.637614] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.638137] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-062d3323-68da-49f3-9eda-d1839184fe55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.647672] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1026.647672] env[61995]: value = "task-795124" [ 1026.647672] env[61995]: _type = "Task" [ 1026.647672] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.660932] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795124, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.682202] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5287e46a-2d75-69f8-c91c-4f3377476a1a, 'name': SearchDatastore_Task, 'duration_secs': 0.029876} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.682491] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.682754] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 410eac71-20cd-4a6d-9b78-e11fa72d74e7/410eac71-20cd-4a6d-9b78-e11fa72d74e7.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1026.683024] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68b81a55-86ab-4606-a413-b9d57157a347 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.691183] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1026.691183] env[61995]: value = "task-795125" [ 1026.691183] env[61995]: _type = "Task" [ 1026.691183] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.708066] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795125, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.711489] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795123, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.782997] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.092559] env[61995]: DEBUG oslo_concurrency.lockutils [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.092559] env[61995]: DEBUG oslo_concurrency.lockutils [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.125372] env[61995]: DEBUG nova.network.neutron [-] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.158356] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795124, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132562} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.158662] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.159950] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f2598a-65f2-4f89-8afd-f56700e8b52b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.189945] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] a48d837d-1008-4248-bbe0-14321368e138/a48d837d-1008-4248-bbe0-14321368e138.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.190317] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0546345e-6704-428e-9f99-c42678a8c37d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.205650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.205650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.205650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.205650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.205813] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.213774] env[61995]: INFO nova.compute.manager [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Terminating instance [ 1027.216969] env[61995]: DEBUG nova.compute.manager [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.217224] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1027.218094] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533b417a-51d2-4fbe-98cf-e2b5f0706bf3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.227429] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1027.227429] env[61995]: value = "task-795126" [ 1027.227429] env[61995]: _type = "Task" [ 1027.227429] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.228023] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795125, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.228325] env[61995]: DEBUG oslo_vmware.api [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795123, 'name': RemoveSnapshot_Task, 'duration_secs': 0.685046} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.228969] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1027.229267] env[61995]: INFO nova.compute.manager [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Took 14.61 seconds to snapshot the instance on the hypervisor. [ 1027.238517] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1027.239180] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8887d8b-2257-40bf-9825-f970519537e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.244202] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-391429bc-ce6c-49eb-907d-f6535e4d1d41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.251338] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795126, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.272108] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance 'd9c53806-68c0-4872-a262-36bc05573674' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1027.283646] env[61995]: DEBUG nova.compute.manager [req-e189252d-4f0c-48a2-bb39-748125cdb765 req-5e438a51-c0ba-4892-b442-75df4897c6d2 service nova] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Received event network-vif-deleted-cdaef75a-1442-4f8d-8eab-340da21a74d5 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.288073] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1027.288073] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.416s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.288387] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.949s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.289846] env[61995]: INFO nova.compute.claims [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1027.327114] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1027.327372] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1027.327630] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleting the datastore file [datastore2] 9e30eadd-5694-4fc7-8b54-2cf1d1571504 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.327917] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10fbcf33-d8e4-4026-bab2-0aed2d05fbdf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.338594] env[61995]: DEBUG oslo_vmware.api [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1027.338594] env[61995]: value = "task-795128" [ 1027.338594] env[61995]: _type = "Task" [ 1027.338594] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.349307] env[61995]: DEBUG oslo_vmware.api [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.595504] env[61995]: INFO nova.compute.manager [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Detaching volume 2453ed83-12b0-456e-9b54-9ba91f0671e0 [ 1027.628599] env[61995]: INFO nova.compute.manager [-] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Took 1.59 seconds to deallocate network for instance. [ 1027.631464] env[61995]: INFO nova.virt.block_device [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Attempting to driver detach volume 2453ed83-12b0-456e-9b54-9ba91f0671e0 from mountpoint /dev/sdb [ 1027.631746] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1027.631963] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185397', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'name': 'volume-2453ed83-12b0-456e-9b54-9ba91f0671e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba89ee63-e293-47e1-90ab-7b8e72dd1b50', 'attached_at': '', 'detached_at': '', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'serial': '2453ed83-12b0-456e-9b54-9ba91f0671e0'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1027.632808] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5717d51-af67-4fe6-a43c-23118a062387 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.660114] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bee735-147d-4a06-ab77-e48ba4da6503 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.668124] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44c73d5-25dd-4704-8237-fc52be58148b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.690805] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f2f248-2767-420b-be16-496fe3303810 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.708452] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.708726] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.710413] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] The volume has not been displaced from its original location: [datastore2] volume-2453ed83-12b0-456e-9b54-9ba91f0671e0/volume-2453ed83-12b0-456e-9b54-9ba91f0671e0.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1027.715725] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Reconfiguring VM instance instance-00000052 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1027.719740] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99c028b9-8c6a-4df3-aba8-3abf2a0c0d8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.740389] env[61995]: DEBUG nova.compute.manager [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Instance disappeared during snapshot {{(pid=61995) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 1027.742290] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795125, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.913172} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.744967] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 410eac71-20cd-4a6d-9b78-e11fa72d74e7/410eac71-20cd-4a6d-9b78-e11fa72d74e7.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1027.745151] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.745323] env[61995]: DEBUG oslo_vmware.api [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1027.745323] env[61995]: value = "task-795129" [ 1027.745323] env[61995]: _type = "Task" [ 1027.745323] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.745530] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc7c5003-1157-42e4-94f9-00bfa9d55e25 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.754158] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795126, 'name': ReconfigVM_Task, 'duration_secs': 0.390452} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.755163] env[61995]: DEBUG nova.compute.manager [None req-29350802-affa-4241-85f1-f730d149fee5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image not found during clean up 17255c28-b10b-4ad7-bafd-de8e81f4c675 {{(pid=61995) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 1027.758880] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Reconfigured VM instance instance-0000005b to attach disk [datastore2] a48d837d-1008-4248-bbe0-14321368e138/a48d837d-1008-4248-bbe0-14321368e138.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.758880] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d681a9b-51e3-4d89-9819-83766802f69c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.760448] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1027.760448] env[61995]: value = "task-795130" [ 1027.760448] env[61995]: _type = "Task" [ 1027.760448] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.767021] env[61995]: DEBUG oslo_vmware.api [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795129, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.772245] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1027.772245] env[61995]: value = "task-795131" [ 1027.772245] env[61995]: _type = "Task" [ 1027.772245] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.776147] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.781070] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1027.781733] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdf223a2-085f-4adf-8d4c-147edcc499e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.789527] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795131, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.791017] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1027.791017] env[61995]: value = "task-795132" [ 1027.791017] env[61995]: _type = "Task" [ 1027.791017] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.805709] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.851576] env[61995]: DEBUG oslo_vmware.api [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.140787] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.217426] env[61995]: DEBUG nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1028.258040] env[61995]: DEBUG oslo_vmware.api [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795129, 'name': ReconfigVM_Task, 'duration_secs': 0.330219} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.258414] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Reconfigured VM instance instance-00000052 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1028.263842] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37a5a067-2f8b-46f2-beb8-eb1f787a887d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.284900] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166941} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.288542] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.288884] env[61995]: DEBUG oslo_vmware.api [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1028.288884] env[61995]: value = "task-795133" [ 1028.288884] env[61995]: _type = "Task" [ 1028.288884] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.289126] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795131, 'name': Rename_Task, 'duration_secs': 0.250555} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.289795] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62005cef-58d6-49ee-9c0f-13d2f0a4c6ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.292128] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1028.295274] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8927457e-6c1a-41cb-a1eb-ddd4fa6b5af1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.321148] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 410eac71-20cd-4a6d-9b78-e11fa72d74e7/410eac71-20cd-4a6d-9b78-e11fa72d74e7.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.328189] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e5da5db-18ff-42bf-9645-83e8bd0d0e29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.344036] env[61995]: DEBUG oslo_vmware.api [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.344453] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1028.344453] env[61995]: value = "task-795134" [ 1028.344453] env[61995]: _type = "Task" [ 1028.344453] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.344726] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795132, 'name': PowerOffVM_Task, 'duration_secs': 0.249268} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.349058] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.349289] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance 'd9c53806-68c0-4872-a262-36bc05573674' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1028.362196] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1028.362196] env[61995]: value = "task-795135" [ 1028.362196] env[61995]: _type = "Task" [ 1028.362196] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.368609] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795134, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.368942] env[61995]: DEBUG oslo_vmware.api [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.625197} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.374564] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.374766] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1028.374950] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1028.375156] env[61995]: INFO nova.compute.manager [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1028.375407] env[61995]: DEBUG oslo.service.loopingcall [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.375903] env[61995]: DEBUG nova.compute.manager [-] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1028.376098] env[61995]: DEBUG nova.network.neutron [-] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1028.384073] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.676389] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e9ba04-94a0-4767-af54-e4887859e102 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.685430] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82107b4-cae7-48da-bf22-87af3ddba240 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.724416] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908cfe3f-34bb-4050-bf56-5c85231593af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.736033] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db73995d-6f03-4b8a-bf66-cb1555b7f8c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.740957] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.752935] env[61995]: DEBUG nova.compute.provider_tree [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.802458] env[61995]: DEBUG oslo_vmware.api [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795133, 'name': ReconfigVM_Task, 'duration_secs': 0.230798} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.802835] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185397', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'name': 'volume-2453ed83-12b0-456e-9b54-9ba91f0671e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba89ee63-e293-47e1-90ab-7b8e72dd1b50', 'attached_at': '', 'detached_at': '', 'volume_id': '2453ed83-12b0-456e-9b54-9ba91f0671e0', 'serial': '2453ed83-12b0-456e-9b54-9ba91f0671e0'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1028.861452] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.861739] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.861904] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.862105] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.862331] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.862411] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.862627] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.862813] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.862957] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.863140] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.863325] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.868257] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795134, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.868613] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-627b7f95-78b4-41d5-8bd2-6261cbb6b783 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.888381] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1028.888381] env[61995]: value = "task-795136" [ 1028.888381] env[61995]: _type = "Task" [ 1028.888381] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.891673] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795135, 'name': ReconfigVM_Task, 'duration_secs': 0.473379} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.895042] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 410eac71-20cd-4a6d-9b78-e11fa72d74e7/410eac71-20cd-4a6d-9b78-e11fa72d74e7.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.895697] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59446733-390e-4285-a210-7b6056dedaff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.902895] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1028.902895] env[61995]: value = "task-795137" [ 1028.902895] env[61995]: _type = "Task" [ 1028.902895] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.905941] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795136, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.917393] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795137, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.193914] env[61995]: DEBUG nova.network.neutron [-] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.259862] env[61995]: DEBUG nova.scheduler.client.report [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.314250] env[61995]: DEBUG nova.compute.manager [req-e5d20e2a-ebfc-428a-ad40-5abbbc09c9d9 req-93ec0a13-d305-461d-b2ed-9dfd7f5b71b2 service nova] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Received event network-vif-deleted-41529e93-5bd1-44e2-9904-dc6702ff0282 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1029.364687] env[61995]: DEBUG oslo_vmware.api [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795134, 'name': PowerOnVM_Task, 'duration_secs': 0.678609} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.365271] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1029.366706] env[61995]: INFO nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Took 9.60 seconds to spawn the instance on the hypervisor. [ 1029.366706] env[61995]: DEBUG nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1029.367578] env[61995]: DEBUG nova.objects.instance [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'flavor' on Instance uuid ba89ee63-e293-47e1-90ab-7b8e72dd1b50 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.369844] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fd2db5-8107-46d3-850f-2ff9227439d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.401985] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795136, 'name': ReconfigVM_Task, 'duration_secs': 0.20772} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.402778] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance 'd9c53806-68c0-4872-a262-36bc05573674' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1029.415789] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795137, 'name': Rename_Task, 'duration_secs': 0.187492} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.416215] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1029.416515] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-795b117b-1fc4-4869-82e7-b0a38c877997 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.424718] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1029.424718] env[61995]: value = "task-795138" [ 1029.424718] env[61995]: _type = "Task" [ 1029.424718] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.433450] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.696573] env[61995]: INFO nova.compute.manager [-] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Took 1.32 seconds to deallocate network for instance. [ 1029.765747] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.766356] env[61995]: DEBUG nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.768914] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.751s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.773018] env[61995]: INFO nova.compute.claims [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1029.889030] env[61995]: INFO nova.compute.manager [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Took 22.13 seconds to build instance. [ 1029.912507] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.912903] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.912903] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.913098] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.913278] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.913456] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.913707] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.913887] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.914102] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.914307] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.914494] env[61995]: DEBUG nova.virt.hardware [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.920128] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Reconfiguring VM instance instance-0000001c to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1029.920728] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed2fa6c5-ad47-42e0-9ada-35aba275f94b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.944549] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795138, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.946146] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1029.946146] env[61995]: value = "task-795139" [ 1029.946146] env[61995]: _type = "Task" [ 1029.946146] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.956184] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795139, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.202869] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.276526] env[61995]: DEBUG nova.compute.utils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.278434] env[61995]: DEBUG nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1030.278602] env[61995]: DEBUG nova.network.neutron [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1030.333688] env[61995]: DEBUG nova.policy [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caf9bbbd848d4d7ba53ba3bd553943e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2018aae65cf4bf98b5ac82519201cf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1030.378046] env[61995]: DEBUG oslo_concurrency.lockutils [None req-effd35b2-f6e7-4332-836d-8163e3c073a8 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.285s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.391234] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f875c7f-ecdb-4915-b09f-acbc737d97c0 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "a48d837d-1008-4248-bbe0-14321368e138" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.640s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.446026] env[61995]: DEBUG oslo_vmware.api [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795138, 'name': PowerOnVM_Task, 'duration_secs': 0.955015} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.446313] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.446522] env[61995]: INFO nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Took 8.34 seconds to spawn the instance on the hypervisor. [ 1030.446704] env[61995]: DEBUG nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.447546] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1de4e3-d5ae-4260-9911-c3234f06df70 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.463199] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795139, 'name': ReconfigVM_Task, 'duration_secs': 0.291747} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.463709] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Reconfigured VM instance instance-0000001c to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1030.464497] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd98c327-be83-4f2c-99d9-73d164f65266 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.487957] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] d9c53806-68c0-4872-a262-36bc05573674/d9c53806-68c0-4872-a262-36bc05573674.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.488310] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13fd4c68-a2dc-4d0b-a47c-fdc5e4e77f51 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.510940] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1030.510940] env[61995]: value = "task-795140" [ 1030.510940] env[61995]: _type = "Task" [ 1030.510940] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.520717] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.684046] env[61995]: DEBUG nova.network.neutron [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Successfully created port: c866f945-735c-48dd-8539-9e28635c0837 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1030.783974] env[61995]: DEBUG nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.974035] env[61995]: INFO nova.compute.manager [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Took 22.68 seconds to build instance. [ 1031.022907] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795140, 'name': ReconfigVM_Task, 'duration_secs': 0.395284} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.025752] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Reconfigured VM instance instance-0000001c to attach disk [datastore2] d9c53806-68c0-4872-a262-36bc05573674/d9c53806-68c0-4872-a262-36bc05573674.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.026254] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance 'd9c53806-68c0-4872-a262-36bc05573674' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1031.106735] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59c2628-6c32-4784-85f1-9d46e08674b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.115590] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07b66b9-0e7f-4ed7-bbf5-fdd153a65e1f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.150282] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090f35b3-5176-4f50-b7e3-59487ad331df {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.158397] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fe5f16-9baf-4a77-bb20-aaf429d6d24e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.173648] env[61995]: DEBUG nova.compute.provider_tree [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.434339] env[61995]: DEBUG nova.compute.manager [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Received event network-changed-505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.434673] env[61995]: DEBUG nova.compute.manager [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Refreshing instance network info cache due to event network-changed-505624a2-00f0-4314-b814-28c5bc7ac5d9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1031.434673] env[61995]: DEBUG oslo_concurrency.lockutils [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] Acquiring lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.434792] env[61995]: DEBUG oslo_concurrency.lockutils [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] Acquired lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.434923] env[61995]: DEBUG nova.network.neutron [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Refreshing network info cache for port 505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.477671] env[61995]: DEBUG oslo_concurrency.lockutils [None req-687e4715-34ef-426d-a09c-0230ea530e67 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.195s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.533369] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c01310-c556-4063-b4bc-fc53ff86d17c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.553118] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed17978-fe36-4561-8e6f-77d1b93a6c1b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.572258] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance 'd9c53806-68c0-4872-a262-36bc05573674' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1031.581303] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.581559] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.581776] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.581965] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.582161] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.584136] env[61995]: INFO nova.compute.manager [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Terminating instance [ 1031.586055] env[61995]: DEBUG nova.compute.manager [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.586259] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1031.587059] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bb8f15-f4e7-4306-adfb-c037a0164d3a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.599233] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1031.599505] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88ce3636-07b0-4a4f-8a7c-feaed0248063 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.607782] env[61995]: DEBUG oslo_vmware.api [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1031.607782] env[61995]: value = "task-795141" [ 1031.607782] env[61995]: _type = "Task" [ 1031.607782] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.619382] env[61995]: DEBUG oslo_vmware.api [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.679915] env[61995]: DEBUG nova.scheduler.client.report [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.798472] env[61995]: DEBUG nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.826637] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.826942] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.827131] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.827333] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.827487] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.827671] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.827903] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.828153] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.828385] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.828564] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.828765] env[61995]: DEBUG nova.virt.hardware [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.830057] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd221be-07f7-47ee-8ebd-f3141f11fe28 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.838959] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e13f6e-6830-4d5b-bd9c-fcf28d89c45a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.096092] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "06886222-5f7f-482e-b5ee-afd7326f2c70" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.096416] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.096610] env[61995]: DEBUG nova.compute.manager [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.097568] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8172e307-965f-49a2-9bd9-be06e13e19dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.105111] env[61995]: DEBUG nova.compute.manager [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1032.105717] env[61995]: DEBUG nova.objects.instance [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lazy-loading 'flavor' on Instance uuid 06886222-5f7f-482e-b5ee-afd7326f2c70 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.115332] env[61995]: DEBUG nova.network.neutron [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Port 802b752a-238e-4b55-8bd3-f73eddc4b0f4 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1032.122153] env[61995]: DEBUG oslo_vmware.api [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795141, 'name': PowerOffVM_Task, 'duration_secs': 0.262746} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.123243] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.123501] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1032.124052] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d86956f9-9b44-478a-9602-4cd094396afc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.185775] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.186424] env[61995]: DEBUG nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1032.189871] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.049s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.190122] env[61995]: DEBUG nova.objects.instance [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'resources' on Instance uuid 453f46c7-5bad-4ca2-b228-f76e62fbd03e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.226112] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1032.226359] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1032.226547] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleting the datastore file [datastore2] ba89ee63-e293-47e1-90ab-7b8e72dd1b50 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.226829] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d34d4a48-5bb8-4cf5-a89b-cbd7987e447a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.240536] env[61995]: DEBUG oslo_vmware.api [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1032.240536] env[61995]: value = "task-795143" [ 1032.240536] env[61995]: _type = "Task" [ 1032.240536] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.255078] env[61995]: DEBUG oslo_vmware.api [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.317289] env[61995]: DEBUG nova.network.neutron [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Updated VIF entry in instance network info cache for port 505624a2-00f0-4314-b814-28c5bc7ac5d9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.317801] env[61995]: DEBUG nova.network.neutron [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Updating instance_info_cache with network_info: [{"id": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "address": "fa:16:3e:52:c6:6f", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap505624a2-00", "ovs_interfaceid": "505624a2-00f0-4314-b814-28c5bc7ac5d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.444403] env[61995]: DEBUG nova.network.neutron [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Successfully updated port: c866f945-735c-48dd-8539-9e28635c0837 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.611040] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.611040] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-551a7add-4337-4ad1-bcac-c3aca0f7734e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.619442] env[61995]: DEBUG oslo_vmware.api [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1032.619442] env[61995]: value = "task-795145" [ 1032.619442] env[61995]: _type = "Task" [ 1032.619442] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.634313] env[61995]: DEBUG oslo_vmware.api [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795145, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.693095] env[61995]: DEBUG nova.compute.utils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1032.698638] env[61995]: DEBUG nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1032.699778] env[61995]: DEBUG nova.network.neutron [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1032.750678] env[61995]: DEBUG nova.policy [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4fa18e6b6fe742909fa15a846d3019e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd23f8abd8f14ec392fbfb7fd5bc64f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1032.756624] env[61995]: DEBUG oslo_vmware.api [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.350133} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.760372] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.760891] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1032.761138] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1032.761414] env[61995]: INFO nova.compute.manager [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1032.761836] env[61995]: DEBUG oslo.service.loopingcall [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.762290] env[61995]: DEBUG nova.compute.manager [-] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1032.762399] env[61995]: DEBUG nova.network.neutron [-] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1032.822391] env[61995]: DEBUG oslo_concurrency.lockutils [req-c2873cdc-6c77-48e8-a0c8-2f40b9aff1b3 req-6582c490-7b95-4b88-a416-eaae271e933c service nova] Releasing lock "refresh_cache-410eac71-20cd-4a6d-9b78-e11fa72d74e7" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.949649] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "refresh_cache-e8bde834-8f55-48d6-8623-1c6e43072c23" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.949649] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "refresh_cache-e8bde834-8f55-48d6-8623-1c6e43072c23" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.949649] env[61995]: DEBUG nova.network.neutron [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1033.026021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3659438-6c99-49c0-8dfa-5061ccb6e996 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.038021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ffdacf-324d-4059-bb1e-81f7b97253e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.070765] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9462eb-8839-4120-a77c-0c8c36805924 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.080301] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d7cf53-4d4b-4f2c-9a6e-990dc815495d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.098233] env[61995]: DEBUG nova.compute.provider_tree [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.134034] env[61995]: DEBUG oslo_vmware.api [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795145, 'name': PowerOffVM_Task, 'duration_secs': 0.256046} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.134034] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1033.134034] env[61995]: DEBUG nova.compute.manager [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.134034] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b3f37a-7288-4719-bad0-63d7f48db976 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.159292] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.159861] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.160226] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.198808] env[61995]: DEBUG nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1033.385652] env[61995]: DEBUG nova.network.neutron [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Successfully created port: c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1033.468441] env[61995]: DEBUG nova.compute.manager [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Received event network-vif-plugged-c866f945-735c-48dd-8539-9e28635c0837 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.468731] env[61995]: DEBUG oslo_concurrency.lockutils [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] Acquiring lock "e8bde834-8f55-48d6-8623-1c6e43072c23-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.468773] env[61995]: DEBUG oslo_concurrency.lockutils [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.468945] env[61995]: DEBUG oslo_concurrency.lockutils [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.470043] env[61995]: DEBUG nova.compute.manager [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] No waiting events found dispatching network-vif-plugged-c866f945-735c-48dd-8539-9e28635c0837 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.470262] env[61995]: WARNING nova.compute.manager [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Received unexpected event network-vif-plugged-c866f945-735c-48dd-8539-9e28635c0837 for instance with vm_state building and task_state spawning. [ 1033.470522] env[61995]: DEBUG nova.compute.manager [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Received event network-changed-c866f945-735c-48dd-8539-9e28635c0837 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.470601] env[61995]: DEBUG nova.compute.manager [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Refreshing instance network info cache due to event network-changed-c866f945-735c-48dd-8539-9e28635c0837. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1033.470768] env[61995]: DEBUG oslo_concurrency.lockutils [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] Acquiring lock "refresh_cache-e8bde834-8f55-48d6-8623-1c6e43072c23" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.484543] env[61995]: DEBUG nova.network.neutron [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1033.606407] env[61995]: DEBUG nova.scheduler.client.report [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.676740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4ade369a-1c26-451e-8ab2-39ed75ae6ea2 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.580s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.854379] env[61995]: DEBUG nova.network.neutron [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Updating instance_info_cache with network_info: [{"id": "c866f945-735c-48dd-8539-9e28635c0837", "address": "fa:16:3e:d2:e2:66", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc866f945-73", "ovs_interfaceid": "c866f945-735c-48dd-8539-9e28635c0837", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.111929] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.922s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.114431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.374s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.116010] env[61995]: INFO nova.compute.claims [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.139775] env[61995]: INFO nova.scheduler.client.report [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted allocations for instance 453f46c7-5bad-4ca2-b228-f76e62fbd03e [ 1034.212027] env[61995]: DEBUG nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1034.217572] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.217769] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.217949] env[61995]: DEBUG nova.network.neutron [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1034.240580] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1034.241061] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1034.241365] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.241699] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1034.242008] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.243297] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1034.243297] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1034.243297] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1034.243480] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1034.243651] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1034.243987] env[61995]: DEBUG nova.virt.hardware [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.245757] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41bf693-b24b-488d-b1b9-0482fa05cff0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.260435] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67091884-e362-4b03-884d-d4ac3792387e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.360026] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "refresh_cache-e8bde834-8f55-48d6-8623-1c6e43072c23" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.360026] env[61995]: DEBUG nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Instance network_info: |[{"id": "c866f945-735c-48dd-8539-9e28635c0837", "address": "fa:16:3e:d2:e2:66", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc866f945-73", "ovs_interfaceid": "c866f945-735c-48dd-8539-9e28635c0837", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1034.360026] env[61995]: DEBUG oslo_concurrency.lockutils [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] Acquired lock "refresh_cache-e8bde834-8f55-48d6-8623-1c6e43072c23" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.360026] env[61995]: DEBUG nova.network.neutron [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Refreshing network info cache for port c866f945-735c-48dd-8539-9e28635c0837 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1034.360026] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:e2:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c866f945-735c-48dd-8539-9e28635c0837', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.368526] env[61995]: DEBUG oslo.service.loopingcall [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.369888] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1034.370299] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-298b090a-6ebb-43fb-8e39-af90fb08d13e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.392186] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.392186] env[61995]: value = "task-795146" [ 1034.392186] env[61995]: _type = "Task" [ 1034.392186] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.400974] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795146, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.401424] env[61995]: DEBUG nova.network.neutron [-] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.648434] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e9da4acb-4cce-4e1f-ac2c-7ad9f2d66c1e tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "453f46c7-5bad-4ca2-b228-f76e62fbd03e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.828s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.787346] env[61995]: DEBUG nova.objects.instance [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lazy-loading 'flavor' on Instance uuid 06886222-5f7f-482e-b5ee-afd7326f2c70 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.903997] env[61995]: INFO nova.compute.manager [-] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Took 2.14 seconds to deallocate network for instance. [ 1034.904285] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795146, 'name': CreateVM_Task, 'duration_secs': 0.358765} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.909708] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1034.910369] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.910567] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.910885] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.916752] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a5a175d-28a8-47b2-b16f-a6996e15809c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.925881] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1034.925881] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52626629-e00d-30d2-26e0-68339cff3bb4" [ 1034.925881] env[61995]: _type = "Task" [ 1034.925881] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.934609] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52626629-e00d-30d2-26e0-68339cff3bb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.958750] env[61995]: DEBUG nova.network.neutron [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.106582] env[61995]: DEBUG nova.network.neutron [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Updated VIF entry in instance network info cache for port c866f945-735c-48dd-8539-9e28635c0837. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1035.107030] env[61995]: DEBUG nova.network.neutron [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Updating instance_info_cache with network_info: [{"id": "c866f945-735c-48dd-8539-9e28635c0837", "address": "fa:16:3e:d2:e2:66", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc866f945-73", "ovs_interfaceid": "c866f945-735c-48dd-8539-9e28635c0837", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.147946] env[61995]: DEBUG nova.scheduler.client.report [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1035.170739] env[61995]: DEBUG nova.scheduler.client.report [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1035.171061] env[61995]: DEBUG nova.compute.provider_tree [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1035.194341] env[61995]: DEBUG nova.scheduler.client.report [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1035.215970] env[61995]: DEBUG nova.scheduler.client.report [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1035.297062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.299972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquired lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.299972] env[61995]: DEBUG nova.network.neutron [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1035.299972] env[61995]: DEBUG nova.objects.instance [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lazy-loading 'info_cache' on Instance uuid 06886222-5f7f-482e-b5ee-afd7326f2c70 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.423431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.428889] env[61995]: DEBUG nova.network.neutron [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Successfully updated port: c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1035.445523] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52626629-e00d-30d2-26e0-68339cff3bb4, 'name': SearchDatastore_Task, 'duration_secs': 0.010305} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.446637] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.447043] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.447615] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.447891] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.448244] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.451635] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca733b70-ad08-497f-b52e-4937e78874b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.468019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.473853] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.474337] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1035.477694] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1502f8f8-2309-40da-8244-1aa5e68a86e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.483881] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1035.483881] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5227ce32-83e1-aad7-1140-0014f314e88b" [ 1035.483881] env[61995]: _type = "Task" [ 1035.483881] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.495281] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5227ce32-83e1-aad7-1140-0014f314e88b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.507414] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26abd8f5-e29f-4e5e-bb41-5640776301ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.514695] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70422e8-2695-447e-ba5b-e733a92d55a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.524076] env[61995]: DEBUG nova.compute.manager [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Received event network-vif-deleted-592fb7d9-e1e2-4233-80f7-6a263f59fac2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.524634] env[61995]: DEBUG nova.compute.manager [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Received event network-vif-plugged-c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.524634] env[61995]: DEBUG oslo_concurrency.lockutils [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] Acquiring lock "83369251-b00f-4595-bf98-28a3ec84f037-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.524901] env[61995]: DEBUG oslo_concurrency.lockutils [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] Lock "83369251-b00f-4595-bf98-28a3ec84f037-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.524901] env[61995]: DEBUG oslo_concurrency.lockutils [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] Lock "83369251-b00f-4595-bf98-28a3ec84f037-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.525022] env[61995]: DEBUG nova.compute.manager [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] No waiting events found dispatching network-vif-plugged-c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.525257] env[61995]: WARNING nova.compute.manager [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Received unexpected event network-vif-plugged-c97b92c2-09c6-42ae-bac7-ad1583a82097 for instance with vm_state building and task_state spawning. [ 1035.525440] env[61995]: DEBUG nova.compute.manager [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Received event network-changed-c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.525601] env[61995]: DEBUG nova.compute.manager [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Refreshing instance network info cache due to event network-changed-c97b92c2-09c6-42ae-bac7-ad1583a82097. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1035.525799] env[61995]: DEBUG oslo_concurrency.lockutils [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] Acquiring lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.526109] env[61995]: DEBUG oslo_concurrency.lockutils [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] Acquired lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.526109] env[61995]: DEBUG nova.network.neutron [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Refreshing network info cache for port c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.555876] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf913c8-44af-49d8-8ad5-64e6e755cdd4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.565427] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eaab8ea-f551-4319-ae5a-7061ea76ce82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.581049] env[61995]: DEBUG nova.compute.provider_tree [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.586952] env[61995]: DEBUG nova.network.neutron [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1035.609725] env[61995]: DEBUG oslo_concurrency.lockutils [req-41466b60-8b59-4889-bf26-48ba93855ad9 req-d4621192-81fb-4c80-a655-e770ddaeec47 service nova] Releasing lock "refresh_cache-e8bde834-8f55-48d6-8623-1c6e43072c23" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.669962] env[61995]: DEBUG nova.network.neutron [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.804810] env[61995]: DEBUG nova.objects.base [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Object Instance<06886222-5f7f-482e-b5ee-afd7326f2c70> lazy-loaded attributes: flavor,info_cache {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1035.935342] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.994378] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5227ce32-83e1-aad7-1140-0014f314e88b, 'name': SearchDatastore_Task, 'duration_secs': 0.020481} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.996313] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c25360-db09-45d5-95fa-f6d7a3e7335d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.998762] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b53dd49-0279-40fe-a77e-ee1735bb51c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.017434] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f6f52d-4774-419e-9091-68675131dac8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.020141] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1036.020141] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52eb8af1-3ced-dcd1-6bc3-776d69af76a7" [ 1036.020141] env[61995]: _type = "Task" [ 1036.020141] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.025641] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance 'd9c53806-68c0-4872-a262-36bc05573674' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1036.034395] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52eb8af1-3ced-dcd1-6bc3-776d69af76a7, 'name': SearchDatastore_Task, 'duration_secs': 0.009677} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.034587] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.034835] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] e8bde834-8f55-48d6-8623-1c6e43072c23/e8bde834-8f55-48d6-8623-1c6e43072c23.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1036.035156] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-918a99f0-668c-46d9-bb79-dcf9029bc980 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.041627] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1036.041627] env[61995]: value = "task-795148" [ 1036.041627] env[61995]: _type = "Task" [ 1036.041627] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.050155] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795148, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.083622] env[61995]: DEBUG nova.scheduler.client.report [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.173057] env[61995]: DEBUG oslo_concurrency.lockutils [req-592dc7fd-0abc-4796-9523-9a7b809e593a req-0b5e0152-4c18-4f68-b84d-0e91e278d3f9 service nova] Releasing lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.173516] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.173797] env[61995]: DEBUG nova.network.neutron [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1036.525474] env[61995]: DEBUG nova.network.neutron [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Updating instance_info_cache with network_info: [{"id": "14a93316-278c-4140-a1e5-567a797d3a44", "address": "fa:16:3e:73:e5:57", "network": {"id": "64d03e8f-2bcf-4c2c-910d-a97bdd1dbb73", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-378669791-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a383c3c42a2f4526ad1fba58b3e00807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14a93316-27", "ovs_interfaceid": "14a93316-278c-4140-a1e5-567a797d3a44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.531814] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1036.532120] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-556076c2-2536-414b-ab97-3a95c021a896 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.539526] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1036.539526] env[61995]: value = "task-795149" [ 1036.539526] env[61995]: _type = "Task" [ 1036.539526] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.549882] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.552811] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795148, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431536} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.553061] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] e8bde834-8f55-48d6-8623-1c6e43072c23/e8bde834-8f55-48d6-8623-1c6e43072c23.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1036.553295] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.553559] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d55e27f-8bf4-4dec-ab69-b39d5dfafb6c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.559658] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1036.559658] env[61995]: value = "task-795150" [ 1036.559658] env[61995]: _type = "Task" [ 1036.559658] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.566667] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795150, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.589120] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.589689] env[61995]: DEBUG nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.592415] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.390s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.592681] env[61995]: DEBUG nova.objects.instance [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lazy-loading 'resources' on Instance uuid 9e30eadd-5694-4fc7-8b54-2cf1d1571504 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.707684] env[61995]: DEBUG nova.network.neutron [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1036.821540] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.821782] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.841420] env[61995]: DEBUG nova.network.neutron [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [{"id": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "address": "fa:16:3e:ae:d5:b8", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc97b92c2-09", "ovs_interfaceid": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.028141] env[61995]: DEBUG oslo_concurrency.lockutils [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Releasing lock "refresh_cache-06886222-5f7f-482e-b5ee-afd7326f2c70" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.049486] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795149, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.067712] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795150, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0643} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.067977] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.068791] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb616fb1-d469-483f-a0b9-4874225ecbde {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.091278] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] e8bde834-8f55-48d6-8623-1c6e43072c23/e8bde834-8f55-48d6-8623-1c6e43072c23.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.091571] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41318d78-7d8e-47e2-8651-16eb1b7ef684 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.108602] env[61995]: DEBUG nova.compute.utils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.110311] env[61995]: DEBUG nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1037.110505] env[61995]: DEBUG nova.network.neutron [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1037.118069] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1037.118069] env[61995]: value = "task-795151" [ 1037.118069] env[61995]: _type = "Task" [ 1037.118069] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.129368] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795151, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.158021] env[61995]: DEBUG nova.policy [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '671c956f9796411e9b66268420075628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26c7cc2bdcb6462d8154127098e94875', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1037.326367] env[61995]: DEBUG nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1037.343791] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.344184] env[61995]: DEBUG nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Instance network_info: |[{"id": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "address": "fa:16:3e:ae:d5:b8", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc97b92c2-09", "ovs_interfaceid": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1037.344784] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:d5:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7894814c-6be3-4b80-a08e-4a771bc05dd1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c97b92c2-09c6-42ae-bac7-ad1583a82097', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.353756] env[61995]: DEBUG oslo.service.loopingcall [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.356588] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1037.357641] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201d9ad9-b392-4f7c-80aa-05a7dc6a5047 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.361375] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b4e9a7c-056f-43ad-863a-415027e6743e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.383809] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e36b491-ca54-4b77-b892-eb9b06da991f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.387697] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.387697] env[61995]: value = "task-795153" [ 1037.387697] env[61995]: _type = "Task" [ 1037.387697] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.419047] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf854f6-2587-40df-b2f1-8c2d340e3b58 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.424573] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795153, 'name': CreateVM_Task} progress is 15%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.429954] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89e1bf3-9b75-4add-ac3f-305cc27e7055 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.444486] env[61995]: DEBUG nova.compute.provider_tree [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.493024] env[61995]: DEBUG nova.network.neutron [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Successfully created port: 913513ba-6ddd-42c3-a202-27978a173b14 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.531684] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1037.532018] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86f6f339-2448-4cc0-a6b0-a445a4fb2271 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.540748] env[61995]: DEBUG oslo_vmware.api [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1037.540748] env[61995]: value = "task-795154" [ 1037.540748] env[61995]: _type = "Task" [ 1037.540748] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.552284] env[61995]: DEBUG oslo_vmware.api [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795154, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.555428] env[61995]: DEBUG oslo_vmware.api [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795149, 'name': PowerOnVM_Task, 'duration_secs': 1.011659} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.555766] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1037.555981] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-fd23c176-f11d-4a45-86f9-910ae8c13b91 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance 'd9c53806-68c0-4872-a262-36bc05573674' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1037.614084] env[61995]: DEBUG nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.628486] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795151, 'name': ReconfigVM_Task, 'duration_secs': 0.316645} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.628903] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Reconfigured VM instance instance-0000005d to attach disk [datastore1] e8bde834-8f55-48d6-8623-1c6e43072c23/e8bde834-8f55-48d6-8623-1c6e43072c23.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.629645] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9e6e7e6-1711-4f8d-b20f-25ef05afb3c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.638114] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1037.638114] env[61995]: value = "task-795155" [ 1037.638114] env[61995]: _type = "Task" [ 1037.638114] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.648024] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795155, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.849192] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.899598] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795153, 'name': CreateVM_Task, 'duration_secs': 0.460313} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.899879] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1037.900894] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.901221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.901739] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.902153] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebd3b780-6ae2-43a1-ae40-e88bae41eb8a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.907338] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1037.907338] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52967227-8aea-0a39-05a0-d14994905df4" [ 1037.907338] env[61995]: _type = "Task" [ 1037.907338] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.914838] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52967227-8aea-0a39-05a0-d14994905df4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.948283] env[61995]: DEBUG nova.scheduler.client.report [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.054720] env[61995]: DEBUG oslo_vmware.api [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795154, 'name': PowerOnVM_Task, 'duration_secs': 0.451757} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.055017] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1038.055214] env[61995]: DEBUG nova.compute.manager [None req-649f7620-22b3-4ad7-b7f4-01f2d854f9f1 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.056069] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06173fc6-669d-4b1b-b44e-738c32fcc343 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.151547] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795155, 'name': Rename_Task, 'duration_secs': 0.209072} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.152901] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1038.153291] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db2be9fc-172c-4c04-92d4-6622445ccd21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.161038] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1038.161038] env[61995]: value = "task-795156" [ 1038.161038] env[61995]: _type = "Task" [ 1038.161038] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.169967] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.417664] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52967227-8aea-0a39-05a0-d14994905df4, 'name': SearchDatastore_Task, 'duration_secs': 0.009169} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.417969] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.418275] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.418528] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.418681] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.418867] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.419154] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98fc26bc-2444-4657-bc9d-96d83bb5cc2d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.427764] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.428021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1038.428775] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e60b206-304f-4095-b8c5-1c4c6eab4853 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.434120] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1038.434120] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d0ae62-8c00-06ad-3210-e147ea7a1453" [ 1038.434120] env[61995]: _type = "Task" [ 1038.434120] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.442715] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d0ae62-8c00-06ad-3210-e147ea7a1453, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.454693] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.457234] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.034s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.457431] env[61995]: DEBUG nova.objects.instance [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'resources' on Instance uuid ba89ee63-e293-47e1-90ab-7b8e72dd1b50 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.478709] env[61995]: INFO nova.scheduler.client.report [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted allocations for instance 9e30eadd-5694-4fc7-8b54-2cf1d1571504 [ 1038.624134] env[61995]: DEBUG nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.649943] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.650248] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.650487] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.650713] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.650866] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.651030] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.651251] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.651416] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.651586] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.651751] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.651928] env[61995]: DEBUG nova.virt.hardware [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.653186] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc68f4f-92d9-4e8e-a942-431f13e31beb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.661183] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0441585-1168-439d-b8cb-33751d5ba10b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.681368] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795156, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.945765] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d0ae62-8c00-06ad-3210-e147ea7a1453, 'name': SearchDatastore_Task, 'duration_secs': 0.01178} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.948563] env[61995]: DEBUG nova.compute.manager [req-1fc04622-08f0-4583-aa25-dcc0e396b640 req-72368404-1721-4a4e-9fea-28efb7c8b855 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Received event network-vif-plugged-913513ba-6ddd-42c3-a202-27978a173b14 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.948771] env[61995]: DEBUG oslo_concurrency.lockutils [req-1fc04622-08f0-4583-aa25-dcc0e396b640 req-72368404-1721-4a4e-9fea-28efb7c8b855 service nova] Acquiring lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.948983] env[61995]: DEBUG oslo_concurrency.lockutils [req-1fc04622-08f0-4583-aa25-dcc0e396b640 req-72368404-1721-4a4e-9fea-28efb7c8b855 service nova] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.949208] env[61995]: DEBUG oslo_concurrency.lockutils [req-1fc04622-08f0-4583-aa25-dcc0e396b640 req-72368404-1721-4a4e-9fea-28efb7c8b855 service nova] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.949414] env[61995]: DEBUG nova.compute.manager [req-1fc04622-08f0-4583-aa25-dcc0e396b640 req-72368404-1721-4a4e-9fea-28efb7c8b855 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] No waiting events found dispatching network-vif-plugged-913513ba-6ddd-42c3-a202-27978a173b14 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1038.949576] env[61995]: WARNING nova.compute.manager [req-1fc04622-08f0-4583-aa25-dcc0e396b640 req-72368404-1721-4a4e-9fea-28efb7c8b855 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Received unexpected event network-vif-plugged-913513ba-6ddd-42c3-a202-27978a173b14 for instance with vm_state building and task_state spawning. [ 1038.949860] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-142a1381-00d1-478f-bd78-04b1ff5bb891 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.956319] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1038.956319] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52baeb84-f111-4338-1af6-18c5e2ce832a" [ 1038.956319] env[61995]: _type = "Task" [ 1038.956319] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.968570] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52baeb84-f111-4338-1af6-18c5e2ce832a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.988173] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5f2c8e41-5982-4c2d-a886-44bde0d9d371 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "9e30eadd-5694-4fc7-8b54-2cf1d1571504" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.781s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.096315] env[61995]: DEBUG nova.network.neutron [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Successfully updated port: 913513ba-6ddd-42c3-a202-27978a173b14 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.177525] env[61995]: DEBUG oslo_vmware.api [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795156, 'name': PowerOnVM_Task, 'duration_secs': 0.871395} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.180026] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1039.180340] env[61995]: INFO nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Took 7.38 seconds to spawn the instance on the hypervisor. [ 1039.180488] env[61995]: DEBUG nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.181477] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443bb44f-1998-40e3-b2be-adf574dbcf17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.221720] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170c7248-a5fd-43e3-8d2e-b255a0437d1b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.228737] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78162693-443d-41f8-a604-7441423a98f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.259552] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f44aab-fc66-413e-a807-ca3ac13c6195 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.267369] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7632b21-3df4-460b-ae63-4d92ee19b9a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.281020] env[61995]: DEBUG nova.compute.provider_tree [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.467156] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52baeb84-f111-4338-1af6-18c5e2ce832a, 'name': SearchDatastore_Task, 'duration_secs': 0.023943} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.467508] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.467831] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1039.468156] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-164b566c-95ef-4842-a229-2d57f1bf9503 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.475320] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1039.475320] env[61995]: value = "task-795158" [ 1039.475320] env[61995]: _type = "Task" [ 1039.475320] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.485237] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795158, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.601971] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "refresh_cache-ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.602208] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "refresh_cache-ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.602429] env[61995]: DEBUG nova.network.neutron [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1039.621845] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1039.622083] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1039.701310] env[61995]: INFO nova.compute.manager [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Took 15.38 seconds to build instance. [ 1039.783981] env[61995]: DEBUG nova.scheduler.client.report [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.918749] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.919067] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.919321] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "e835ac53-29af-4bd0-b186-5c6270ccf760-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.919522] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.919704] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.922021] env[61995]: INFO nova.compute.manager [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Terminating instance [ 1039.924253] env[61995]: DEBUG nova.compute.manager [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1039.924521] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1039.925544] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce721074-6387-4d6b-8af8-e1615bd33aee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.934275] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1039.934570] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8aace768-e69b-482e-a342-6eb958310c75 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.942042] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1039.942042] env[61995]: value = "task-795159" [ 1039.942042] env[61995]: _type = "Task" [ 1039.942042] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.951074] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.987216] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795158, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489903} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.987549] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1039.987776] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.988104] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86a41b23-809d-4a5c-ac9f-e88012c0d6e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.994423] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1039.994423] env[61995]: value = "task-795160" [ 1039.994423] env[61995]: _type = "Task" [ 1039.994423] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.003071] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795160, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.128905] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1040.129786] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1040.129786] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1040.138726] env[61995]: DEBUG nova.network.neutron [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1040.203871] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a7d5e8b8-8b37-4187-91d3-29cd27346c41 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.897s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.290930] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.294479] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.445s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.296204] env[61995]: INFO nova.compute.claims [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.308196] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.309022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.309022] env[61995]: DEBUG nova.compute.manager [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Going to confirm migration 6 {{(pid=61995) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1040.327695] env[61995]: INFO nova.scheduler.client.report [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted allocations for instance ba89ee63-e293-47e1-90ab-7b8e72dd1b50 [ 1040.421991] env[61995]: DEBUG nova.network.neutron [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Updating instance_info_cache with network_info: [{"id": "913513ba-6ddd-42c3-a202-27978a173b14", "address": "fa:16:3e:5a:9a:94", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap913513ba-6d", "ovs_interfaceid": "913513ba-6ddd-42c3-a202-27978a173b14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.453668] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.505498] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795160, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123872} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.505824] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.506587] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ec9342-c915-4d89-8ac6-94d4c77bf05c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.529150] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.529551] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49bcd076-595e-43df-a3b5-90abcd1596b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.549877] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1040.549877] env[61995]: value = "task-795161" [ 1040.549877] env[61995]: _type = "Task" [ 1040.549877] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.558324] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.634487] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Skipping network cache update for instance because it is being deleted. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1040.634808] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1040.635016] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1040.683050] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.683275] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.683439] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1040.683600] env[61995]: DEBUG nova.objects.instance [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lazy-loading 'info_cache' on Instance uuid d9c53806-68c0-4872-a262-36bc05573674 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.844450] env[61995]: DEBUG oslo_concurrency.lockutils [None req-7b7f89e7-87ff-46f1-84e4-2ccfadbbbf66 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "ba89ee63-e293-47e1-90ab-7b8e72dd1b50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.263s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.881198] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.925461] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "refresh_cache-ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.926149] env[61995]: DEBUG nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Instance network_info: |[{"id": "913513ba-6ddd-42c3-a202-27978a173b14", "address": "fa:16:3e:5a:9a:94", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap913513ba-6d", "ovs_interfaceid": "913513ba-6ddd-42c3-a202-27978a173b14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1040.927133] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:9a:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '913513ba-6ddd-42c3-a202-27978a173b14', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.934842] env[61995]: DEBUG oslo.service.loopingcall [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.935551] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1040.935812] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-497c572a-71b0-451b-9158-c14b5bf5b767 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.964716] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795159, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.966654] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.966654] env[61995]: value = "task-795162" [ 1040.966654] env[61995]: _type = "Task" [ 1040.966654] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.976765] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795162, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.032952] env[61995]: DEBUG nova.compute.manager [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Received event network-changed-913513ba-6ddd-42c3-a202-27978a173b14 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1041.033195] env[61995]: DEBUG nova.compute.manager [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Refreshing instance network info cache due to event network-changed-913513ba-6ddd-42c3-a202-27978a173b14. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1041.033435] env[61995]: DEBUG oslo_concurrency.lockutils [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] Acquiring lock "refresh_cache-ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.033590] env[61995]: DEBUG oslo_concurrency.lockutils [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] Acquired lock "refresh_cache-ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.033756] env[61995]: DEBUG nova.network.neutron [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Refreshing network info cache for port 913513ba-6ddd-42c3-a202-27978a173b14 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1041.063959] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795161, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.462144] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "e8bde834-8f55-48d6-8623-1c6e43072c23" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.462347] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.462466] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "e8bde834-8f55-48d6-8623-1c6e43072c23-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.462659] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.462834] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.464793] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795159, 'name': PowerOffVM_Task, 'duration_secs': 1.110715} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.467646] env[61995]: INFO nova.compute.manager [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Terminating instance [ 1041.469383] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1041.469722] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1041.470366] env[61995]: DEBUG nova.compute.manager [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1041.470553] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1041.470999] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9532efaf-c99d-4d06-b6d1-f038a7503f4b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.476490] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3668583c-c909-44ec-9086-f179597afed3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.487652] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1041.487875] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795162, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.488125] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5eb62b1-2214-4372-ae02-34bbeb14fe99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.495775] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1041.495775] env[61995]: value = "task-795164" [ 1041.495775] env[61995]: _type = "Task" [ 1041.495775] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.498523] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "a48d837d-1008-4248-bbe0-14321368e138" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.498757] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "a48d837d-1008-4248-bbe0-14321368e138" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.499071] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "a48d837d-1008-4248-bbe0-14321368e138-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.499295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "a48d837d-1008-4248-bbe0-14321368e138-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.499473] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "a48d837d-1008-4248-bbe0-14321368e138-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.505373] env[61995]: INFO nova.compute.manager [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Terminating instance [ 1041.507158] env[61995]: DEBUG nova.compute.manager [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1041.507376] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1041.508196] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d8e31b-7287-410f-8fd5-761503480f12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.514289] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.521692] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1041.521692] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46f91ec8-9288-4eaa-bb15-2cc07787d2e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.528166] env[61995]: DEBUG oslo_vmware.api [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1041.528166] env[61995]: value = "task-795165" [ 1041.528166] env[61995]: _type = "Task" [ 1041.528166] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.534821] env[61995]: DEBUG oslo_vmware.api [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.558787] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1041.559129] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1041.559397] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Deleting the datastore file [datastore1] e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.559742] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6829554-fd41-42ed-9a12-0320aecab053 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.567787] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795161, 'name': ReconfigVM_Task, 'duration_secs': 0.696096} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.570484] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.571253] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1041.571253] env[61995]: value = "task-795167" [ 1041.571253] env[61995]: _type = "Task" [ 1041.571253] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.572044] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70441c6-f95a-442c-b6f2-b792a2a24c51 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.574642] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-34585754-40e3-48c1-ba51-b2ccf908d92b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.585332] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795167, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.588182] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1041.588182] env[61995]: value = "task-795168" [ 1041.588182] env[61995]: _type = "Task" [ 1041.588182] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.589300] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55136c8a-33bb-460a-bcdc-999a9714582c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.634469] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dedd07a-f6ea-4566-aff0-b5007c18be79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.638961] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795168, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.650207] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d439a727-69e2-4c33-8c9d-dee44aacf499 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.662890] env[61995]: DEBUG nova.compute.provider_tree [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.880033] env[61995]: DEBUG nova.network.neutron [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Updated VIF entry in instance network info cache for port 913513ba-6ddd-42c3-a202-27978a173b14. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1041.880594] env[61995]: DEBUG nova.network.neutron [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Updating instance_info_cache with network_info: [{"id": "913513ba-6ddd-42c3-a202-27978a173b14", "address": "fa:16:3e:5a:9a:94", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap913513ba-6d", "ovs_interfaceid": "913513ba-6ddd-42c3-a202-27978a173b14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.982787] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795162, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.008208] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.036310] env[61995]: DEBUG oslo_vmware.api [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795165, 'name': PowerOffVM_Task, 'duration_secs': 0.32802} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.036584] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1042.036768] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1042.037032] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8cbdb826-fd7d-4422-9009-3472b244b221 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.086840] env[61995]: DEBUG oslo_vmware.api [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795167, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16697} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.086840] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.086997] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1042.089351] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1042.089351] env[61995]: INFO nova.compute.manager [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Took 2.16 seconds to destroy the instance on the hypervisor. [ 1042.089351] env[61995]: DEBUG oslo.service.loopingcall [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.089351] env[61995]: DEBUG nova.compute.manager [-] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.089351] env[61995]: DEBUG nova.network.neutron [-] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1042.100697] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795168, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.168875] env[61995]: DEBUG nova.scheduler.client.report [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.185241] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1042.185729] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1042.186052] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleting the datastore file [datastore2] a48d837d-1008-4248-bbe0-14321368e138 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.186465] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5096c8a-0ae4-43bc-b269-5ac7dfd5f088 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.193499] env[61995]: DEBUG oslo_vmware.api [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1042.193499] env[61995]: value = "task-795170" [ 1042.193499] env[61995]: _type = "Task" [ 1042.193499] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.207746] env[61995]: DEBUG oslo_vmware.api [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.383178] env[61995]: DEBUG oslo_concurrency.lockutils [req-e37adce9-4a64-4e57-8587-daa74e3e2a76 req-4d5560a1-75f3-4ff2-b013-cab354fd37d9 service nova] Releasing lock "refresh_cache-ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.481672] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795162, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.508950] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.534885] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.603335] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795168, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.675516] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.675516] env[61995]: DEBUG nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.706175] env[61995]: DEBUG oslo_vmware.api [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147775} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.706175] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.706175] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1042.706175] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1042.706175] env[61995]: INFO nova.compute.manager [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: a48d837d-1008-4248-bbe0-14321368e138] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1042.706175] env[61995]: DEBUG oslo.service.loopingcall [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.706175] env[61995]: DEBUG nova.compute.manager [-] [instance: a48d837d-1008-4248-bbe0-14321368e138] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.706175] env[61995]: DEBUG nova.network.neutron [-] [instance: a48d837d-1008-4248-bbe0-14321368e138] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1042.957472] env[61995]: DEBUG nova.compute.manager [req-1eddecf0-60f3-44c1-8f38-47ca9aac95fc req-1b12a342-cbf1-4680-a784-9d6edca0784b service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Received event network-vif-deleted-97857bcc-fdda-481c-a35a-3c3c3ec6d995 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.957739] env[61995]: INFO nova.compute.manager [req-1eddecf0-60f3-44c1-8f38-47ca9aac95fc req-1b12a342-cbf1-4680-a784-9d6edca0784b service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Neutron deleted interface 97857bcc-fdda-481c-a35a-3c3c3ec6d995; detaching it from the instance and deleting it from the info cache [ 1042.957959] env[61995]: DEBUG nova.network.neutron [req-1eddecf0-60f3-44c1-8f38-47ca9aac95fc req-1b12a342-cbf1-4680-a784-9d6edca0784b service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.981607] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795162, 'name': CreateVM_Task, 'duration_secs': 1.56765} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.981790] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1042.982528] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.982735] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.983159] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1042.983408] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee69d7c1-ff6b-4a02-8cd1-7c41fe9e9c3d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.987982] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1042.987982] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523f109d-83b3-5e0f-2bb9-d35a6c2e9a0e" [ 1042.987982] env[61995]: _type = "Task" [ 1042.987982] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.995894] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523f109d-83b3-5e0f-2bb9-d35a6c2e9a0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.008698] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795164, 'name': PowerOffVM_Task, 'duration_secs': 1.22824} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.008922] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1043.009138] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1043.009393] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3464ebcb-0078-47af-92d5-0fdbf8643dac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.038104] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.038452] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1043.038768] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.039020] env[61995]: DEBUG nova.network.neutron [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.039270] env[61995]: DEBUG nova.objects.instance [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'info_cache' on Instance uuid d9c53806-68c0-4872-a262-36bc05573674 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.041207] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.041668] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.041839] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.041996] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.042161] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.042310] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.042444] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1043.042585] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.059915] env[61995]: DEBUG nova.compute.manager [req-68d07388-bee7-420f-845c-dca161463d1b req-41a6ffa8-8180-42c9-bda3-3c9e4bb4a783 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Received event network-vif-deleted-8c280b5d-bf84-4a8d-8ab7-4230645c59d1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1043.060233] env[61995]: INFO nova.compute.manager [req-68d07388-bee7-420f-845c-dca161463d1b req-41a6ffa8-8180-42c9-bda3-3c9e4bb4a783 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Neutron deleted interface 8c280b5d-bf84-4a8d-8ab7-4230645c59d1; detaching it from the instance and deleting it from the info cache [ 1043.060415] env[61995]: DEBUG nova.network.neutron [req-68d07388-bee7-420f-845c-dca161463d1b req-41a6ffa8-8180-42c9-bda3-3c9e4bb4a783 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.097023] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1043.097023] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1043.097023] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleting the datastore file [datastore1] e8bde834-8f55-48d6-8623-1c6e43072c23 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.099441] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5973ca59-05db-486b-aaf0-7eb3986d1f65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.106323] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795168, 'name': Rename_Task, 'duration_secs': 1.111845} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.107607] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1043.107929] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1043.107929] env[61995]: value = "task-795172" [ 1043.107929] env[61995]: _type = "Task" [ 1043.107929] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.108186] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94e1901b-2679-46b9-b840-d75694af05db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.117860] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.119101] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1043.119101] env[61995]: value = "task-795173" [ 1043.119101] env[61995]: _type = "Task" [ 1043.119101] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.133223] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795173, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.181102] env[61995]: DEBUG nova.compute.utils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.182558] env[61995]: DEBUG nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1043.182755] env[61995]: DEBUG nova.network.neutron [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1043.249383] env[61995]: DEBUG nova.policy [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1043.429578] env[61995]: DEBUG nova.network.neutron [-] [instance: a48d837d-1008-4248-bbe0-14321368e138] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.446030] env[61995]: DEBUG nova.network.neutron [-] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.462495] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ccaace06-ed18-48b2-81c8-9016d8ccab6a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.472289] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a03e332-9cf0-48d1-96b2-cd4abcad8fb5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.506458] env[61995]: DEBUG nova.compute.manager [req-1eddecf0-60f3-44c1-8f38-47ca9aac95fc req-1b12a342-cbf1-4680-a784-9d6edca0784b service nova] [instance: a48d837d-1008-4248-bbe0-14321368e138] Detach interface failed, port_id=97857bcc-fdda-481c-a35a-3c3c3ec6d995, reason: Instance a48d837d-1008-4248-bbe0-14321368e138 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1043.511333] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523f109d-83b3-5e0f-2bb9-d35a6c2e9a0e, 'name': SearchDatastore_Task, 'duration_secs': 0.009697} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.511820] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.512115] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.512419] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.512629] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.512857] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.513196] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34701228-40c1-42cb-8fd9-25961b7b6569 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.521823] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.522037] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1043.523126] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57813907-d2de-4672-aff7-4fcece3edf76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.528846] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1043.528846] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523b817f-8585-dafa-cf53-11110a2d2a04" [ 1043.528846] env[61995]: _type = "Task" [ 1043.528846] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.539569] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523b817f-8585-dafa-cf53-11110a2d2a04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.548433] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.548665] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.548828] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.548976] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1043.550181] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33198f6-aab9-44b5-84fb-3b00da3ed3f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.560418] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df70c48d-acfa-4c3e-8f82-a756934ff018 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.565476] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2627fb9e-394a-4917-af79-178fcd3871a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.581471] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112c85dd-fbac-4cfe-84f6-10a0fbcc6b94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.586712] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6e03b9-5d81-4bfa-972e-1445add6a791 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.601845] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692f0850-4934-4f89-a416-e2a2c62d713f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.619534] env[61995]: DEBUG nova.compute.manager [req-68d07388-bee7-420f-845c-dca161463d1b req-41a6ffa8-8180-42c9-bda3-3c9e4bb4a783 service nova] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Detach interface failed, port_id=8c280b5d-bf84-4a8d-8ab7-4230645c59d1, reason: Instance e835ac53-29af-4bd0-b186-5c6270ccf760 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1043.646945] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180160MB free_disk=66GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1043.647156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.647396] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.659890] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795173, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.663119] env[61995]: DEBUG oslo_vmware.api [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177016} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.663119] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.663293] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1043.664265] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1043.664265] env[61995]: INFO nova.compute.manager [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Took 2.19 seconds to destroy the instance on the hypervisor. [ 1043.664265] env[61995]: DEBUG oslo.service.loopingcall [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.664265] env[61995]: DEBUG nova.compute.manager [-] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1043.664265] env[61995]: DEBUG nova.network.neutron [-] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1043.685756] env[61995]: DEBUG nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.703977] env[61995]: DEBUG nova.network.neutron [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Successfully created port: 7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.932106] env[61995]: INFO nova.compute.manager [-] [instance: a48d837d-1008-4248-bbe0-14321368e138] Took 1.23 seconds to deallocate network for instance. [ 1043.945153] env[61995]: INFO nova.compute.manager [-] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Took 1.86 seconds to deallocate network for instance. [ 1044.039277] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523b817f-8585-dafa-cf53-11110a2d2a04, 'name': SearchDatastore_Task, 'duration_secs': 0.009235} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.040338] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19ed2810-2e65-4a83-b5d2-27e246385a36 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.046059] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1044.046059] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5288962f-5346-95de-667c-2b8e4bb4f06e" [ 1044.046059] env[61995]: _type = "Task" [ 1044.046059] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.058330] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5288962f-5346-95de-667c-2b8e4bb4f06e, 'name': SearchDatastore_Task, 'duration_secs': 0.009063} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.058597] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.059604] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] ff07d5b8-38b4-48d4-a6d0-632d37d3a91a/ff07d5b8-38b4-48d4-a6d0-632d37d3a91a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1044.059604] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2ee669c-bd62-4864-9021-2af936bf3908 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.065845] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1044.065845] env[61995]: value = "task-795175" [ 1044.065845] env[61995]: _type = "Task" [ 1044.065845] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.074648] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.093544] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.093841] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.166263] env[61995]: DEBUG oslo_vmware.api [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795173, 'name': PowerOnVM_Task, 'duration_secs': 0.718388} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.167241] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1044.167446] env[61995]: INFO nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Took 9.96 seconds to spawn the instance on the hypervisor. [ 1044.167919] env[61995]: DEBUG nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.168543] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cdbc1e3-17b1-47ea-91a7-36a117a03119 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.308729] env[61995]: DEBUG nova.network.neutron [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [{"id": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "address": "fa:16:3e:b2:a7:5b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap802b752a-23", "ovs_interfaceid": "802b752a-238e-4b55-8bd3-f73eddc4b0f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.439029] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.449835] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.500019] env[61995]: DEBUG nova.network.neutron [-] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.576542] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795175, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.597235] env[61995]: DEBUG nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1044.668355] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Applying migration context for instance d9c53806-68c0-4872-a262-36bc05573674 as it has an incoming, in-progress migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2. Migration status is confirming {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1044.670599] env[61995]: INFO nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating resource usage from migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2 [ 1044.687604] env[61995]: INFO nova.compute.manager [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Took 19.68 seconds to build instance. [ 1044.692781] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b40b1866-ced3-40ef-9ab7-30d58b75f288 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.692923] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ed952baf-f1ef-4ec0-8a15-928287190456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.693061] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e835ac53-29af-4bd0-b186-5c6270ccf760 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.693181] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 24b974c0-9567-46a3-96f6-fac1e602f46e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 5cb4c1d9-e661-48b4-8600-1b823d63ddbb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 66c56f7b-1fc7-4aed-8afc-350817e1ca48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 06886222-5f7f-482e-b5ee-afd7326f2c70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 432f4d05-ea4d-402c-8934-3aa1aaf93d48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance a48d837d-1008-4248-bbe0-14321368e138 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 410eac71-20cd-4a6d-9b78-e11fa72d74e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1044.694175] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d9c53806-68c0-4872-a262-36bc05573674 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694616] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e8bde834-8f55-48d6-8623-1c6e43072c23 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694616] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 83369251-b00f-4595-bf98-28a3ec84f037 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694616] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ff07d5b8-38b4-48d4-a6d0-632d37d3a91a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.694616] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance c3b2e000-cfa8-4b5b-8763-982c6b768621 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.699255] env[61995]: DEBUG nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.726121] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.727115] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.727350] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.727567] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.727726] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.727882] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.728165] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.728367] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.728553] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.728723] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.728907] env[61995]: DEBUG nova.virt.hardware [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.729851] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6abed6d-6f51-4839-8f62-137879ae15b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.737813] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8109e7-c3bf-4066-8530-c669497bfd41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.811507] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-d9c53806-68c0-4872-a262-36bc05573674" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.811848] env[61995]: DEBUG nova.objects.instance [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'migration_context' on Instance uuid d9c53806-68c0-4872-a262-36bc05573674 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.003430] env[61995]: INFO nova.compute.manager [-] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Took 1.34 seconds to deallocate network for instance. [ 1045.028140] env[61995]: DEBUG nova.compute.manager [req-f6b2597b-5060-4f45-b537-e975ed88ca0c req-12ef026d-4626-4663-9671-003c144a7416 service nova] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Received event network-vif-deleted-c866f945-735c-48dd-8539-9e28635c0837 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1045.077275] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546497} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.077275] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] ff07d5b8-38b4-48d4-a6d0-632d37d3a91a/ff07d5b8-38b4-48d4-a6d0-632d37d3a91a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1045.077732] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.078856] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-773e4dbe-302d-4d24-8d73-d8ddbd497c9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.088268] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1045.088268] env[61995]: value = "task-795176" [ 1045.088268] env[61995]: _type = "Task" [ 1045.088268] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.097708] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795176, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.117381] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.189773] env[61995]: DEBUG oslo_concurrency.lockutils [None req-adca4919-e441-4ac2-a999-9f113d56e64c tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.195s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.200019] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e2014437-fbcd-454a-893f-9ad7c7d461d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1045.200019] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1045.200019] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3712MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1045.315922] env[61995]: DEBUG nova.objects.base [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1045.317379] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f7d86d-8586-4506-818a-a4342adfdead {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.344451] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b21a451-7019-443c-a4a8-60fb074132d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.353838] env[61995]: DEBUG oslo_vmware.api [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1045.353838] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521123e6-866f-0739-11a3-9fdc7fd13788" [ 1045.353838] env[61995]: _type = "Task" [ 1045.353838] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.363419] env[61995]: DEBUG oslo_vmware.api [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521123e6-866f-0739-11a3-9fdc7fd13788, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.467440] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35334be6-9ec2-4f25-8feb-5c4dd1d3873f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.475218] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f437a6-1baf-43b6-a46b-69df2516520e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.520364] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.521751] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a9fb7f-2625-4c31-b854-9a902166d1b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.530432] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c646cbb-f6b6-45b9-9c86-2254d0fa2d38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.544478] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.598514] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083658} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.598791] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.599682] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d999e3-68cc-41f3-bdaf-e9f8f63a2750 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.622792] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] ff07d5b8-38b4-48d4-a6d0-632d37d3a91a/ff07d5b8-38b4-48d4-a6d0-632d37d3a91a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.628108] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e27c4d19-2ef8-48b1-a053-a1c1c8b9783f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.644157] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1045.644157] env[61995]: value = "task-795178" [ 1045.644157] env[61995]: _type = "Task" [ 1045.644157] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.653982] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795178, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.771494] env[61995]: DEBUG nova.compute.manager [req-7dfb7908-bbd8-4bb8-8148-646cae96fafa req-ac5c4dec-cfaf-4d7b-83d1-7ce4e83f7091 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-vif-plugged-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1045.771494] env[61995]: DEBUG oslo_concurrency.lockutils [req-7dfb7908-bbd8-4bb8-8148-646cae96fafa req-ac5c4dec-cfaf-4d7b-83d1-7ce4e83f7091 service nova] Acquiring lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.771494] env[61995]: DEBUG oslo_concurrency.lockutils [req-7dfb7908-bbd8-4bb8-8148-646cae96fafa req-ac5c4dec-cfaf-4d7b-83d1-7ce4e83f7091 service nova] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.771494] env[61995]: DEBUG oslo_concurrency.lockutils [req-7dfb7908-bbd8-4bb8-8148-646cae96fafa req-ac5c4dec-cfaf-4d7b-83d1-7ce4e83f7091 service nova] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.771494] env[61995]: DEBUG nova.compute.manager [req-7dfb7908-bbd8-4bb8-8148-646cae96fafa req-ac5c4dec-cfaf-4d7b-83d1-7ce4e83f7091 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] No waiting events found dispatching network-vif-plugged-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.771494] env[61995]: WARNING nova.compute.manager [req-7dfb7908-bbd8-4bb8-8148-646cae96fafa req-ac5c4dec-cfaf-4d7b-83d1-7ce4e83f7091 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received unexpected event network-vif-plugged-7efb2852-0c7e-4e86-8996-8cba798b76b2 for instance with vm_state building and task_state spawning. [ 1045.864791] env[61995]: DEBUG oslo_vmware.api [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521123e6-866f-0739-11a3-9fdc7fd13788, 'name': SearchDatastore_Task, 'duration_secs': 0.008837} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.865080] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.966758] env[61995]: DEBUG nova.compute.manager [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Received event network-changed-c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1045.966950] env[61995]: DEBUG nova.compute.manager [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Refreshing instance network info cache due to event network-changed-c97b92c2-09c6-42ae-bac7-ad1583a82097. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1045.967111] env[61995]: DEBUG oslo_concurrency.lockutils [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] Acquiring lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.967272] env[61995]: DEBUG oslo_concurrency.lockutils [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] Acquired lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.967433] env[61995]: DEBUG nova.network.neutron [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Refreshing network info cache for port c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1046.047105] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.155145] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.344248] env[61995]: DEBUG nova.network.neutron [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Successfully updated port: 7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.552367] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1046.552665] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.905s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.552923] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.114s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.553390] env[61995]: DEBUG nova.objects.instance [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lazy-loading 'resources' on Instance uuid a48d837d-1008-4248-bbe0-14321368e138 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.655170] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795178, 'name': ReconfigVM_Task, 'duration_secs': 0.803598} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.655439] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Reconfigured VM instance instance-0000005f to attach disk [datastore1] ff07d5b8-38b4-48d4-a6d0-632d37d3a91a/ff07d5b8-38b4-48d4-a6d0-632d37d3a91a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.656093] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b02a8f8-0eee-4a3f-b3b9-962e97b7567a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.663061] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1046.663061] env[61995]: value = "task-795179" [ 1046.663061] env[61995]: _type = "Task" [ 1046.663061] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.676918] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795179, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.719479] env[61995]: DEBUG nova.network.neutron [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updated VIF entry in instance network info cache for port c97b92c2-09c6-42ae-bac7-ad1583a82097. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.720018] env[61995]: DEBUG nova.network.neutron [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [{"id": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "address": "fa:16:3e:ae:d5:b8", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc97b92c2-09", "ovs_interfaceid": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.848561] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.848723] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.848879] env[61995]: DEBUG nova.network.neutron [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.175922] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795179, 'name': Rename_Task, 'duration_secs': 0.178194} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.176222] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1047.176389] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39e9f438-241b-4e44-bbbf-7ea101520c7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.182782] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1047.182782] env[61995]: value = "task-795180" [ 1047.182782] env[61995]: _type = "Task" [ 1047.182782] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.190505] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.223261] env[61995]: DEBUG oslo_concurrency.lockutils [req-0e07f96f-8b77-46e2-aa02-d71bb7987893 req-0c90e5b7-a516-4a09-962d-625ed6650b97 service nova] Releasing lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.272548] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2a40a0-f5bb-4a8c-9693-22329169f044 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.281416] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24889510-eb32-43ed-a119-714976537b7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.311503] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec02d3f-4335-41a4-afae-b7ec4135d7fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.318622] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bf8cec-7126-4ca5-a92d-874180be1e1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.331347] env[61995]: DEBUG nova.compute.provider_tree [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.381015] env[61995]: DEBUG nova.network.neutron [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1047.511107] env[61995]: DEBUG nova.network.neutron [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.692484] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795180, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.834356] env[61995]: DEBUG nova.scheduler.client.report [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.992172] env[61995]: DEBUG nova.compute.manager [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1047.992419] env[61995]: DEBUG nova.compute.manager [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing instance network info cache due to event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1047.992645] env[61995]: DEBUG oslo_concurrency.lockutils [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.013461] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.013777] env[61995]: DEBUG nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Instance network_info: |[{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1048.014117] env[61995]: DEBUG oslo_concurrency.lockutils [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.014351] env[61995]: DEBUG nova.network.neutron [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.015664] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:3b:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7efb2852-0c7e-4e86-8996-8cba798b76b2', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1048.023236] env[61995]: DEBUG oslo.service.loopingcall [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.026218] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1048.028694] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36bc66a5-46fe-4222-9864-f13ec52bdab6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.048824] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1048.048824] env[61995]: value = "task-795182" [ 1048.048824] env[61995]: _type = "Task" [ 1048.048824] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.057432] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795182, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.194517] env[61995]: DEBUG oslo_vmware.api [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795180, 'name': PowerOnVM_Task, 'duration_secs': 0.90426} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.194877] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1048.196163] env[61995]: INFO nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Took 9.57 seconds to spawn the instance on the hypervisor. [ 1048.196163] env[61995]: DEBUG nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.196344] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a7420f-16f5-4e76-a1c5-c6d69778c4be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.243014] env[61995]: DEBUG nova.network.neutron [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updated VIF entry in instance network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1048.243605] env[61995]: DEBUG nova.network.neutron [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.340072] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.343283] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.894s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.343794] env[61995]: DEBUG nova.objects.instance [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'resources' on Instance uuid e835ac53-29af-4bd0-b186-5c6270ccf760 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.367277] env[61995]: INFO nova.scheduler.client.report [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleted allocations for instance a48d837d-1008-4248-bbe0-14321368e138 [ 1048.558710] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795182, 'name': CreateVM_Task, 'duration_secs': 0.315748} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.558886] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1048.559638] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.559805] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.560153] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.560423] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dec2eef3-d205-4a65-88f7-ea4e68ce4736 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.564693] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1048.564693] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526d5888-aac0-5e21-c6ec-a38b9579b000" [ 1048.564693] env[61995]: _type = "Task" [ 1048.564693] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.572319] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526d5888-aac0-5e21-c6ec-a38b9579b000, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.714076] env[61995]: INFO nova.compute.manager [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Took 19.99 seconds to build instance. [ 1048.747028] env[61995]: DEBUG oslo_concurrency.lockutils [req-7c478e42-6a51-4ff5-9d16-32a84b6a64d2 req-7592f4ad-b408-44a7-8a32-691d1488dbe8 service nova] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.876032] env[61995]: DEBUG oslo_concurrency.lockutils [None req-54e385a9-546b-4ff8-ab52-a48d229e2c6d tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "a48d837d-1008-4248-bbe0-14321368e138" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.377s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.042333] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eaaf821-4c59-4b0a-83c3-2c278e1efc69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.049884] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818e6656-b942-4018-a8ec-7e0c53f7abcd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.082939] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171c88b3-10ea-465a-8b13-7cf5efc4e7fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.093391] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcc8ec3-41e3-492f-8144-0095c93b06a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.097076] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526d5888-aac0-5e21-c6ec-a38b9579b000, 'name': SearchDatastore_Task, 'duration_secs': 0.008907} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.097731] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.097968] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1049.098275] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.098427] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.098704] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1049.099226] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22346137-1250-4283-8c46-b2fded403767 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.108415] env[61995]: DEBUG nova.compute.provider_tree [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.115917] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1049.115917] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1049.116492] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18ef6e0a-61d6-4a63-9871-89bea5ac4d31 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.121849] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1049.121849] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f4f0d8-7ff6-a818-4de3-d1296a1d320c" [ 1049.121849] env[61995]: _type = "Task" [ 1049.121849] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.129683] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f4f0d8-7ff6-a818-4de3-d1296a1d320c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.216702] env[61995]: DEBUG oslo_concurrency.lockutils [None req-450fcf56-a250-4dac-9464-a34f605589b6 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.508s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.382639] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670f239a-1507-4906-8b61-c3023eec0eb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.389570] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5a7ffd-44bd-4844-81b2-4534c3732094 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Suspending the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1049.389819] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5a8f830a-3595-44ee-851f-692bf64d15da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.396853] env[61995]: DEBUG oslo_vmware.api [None req-2b5a7ffd-44bd-4844-81b2-4534c3732094 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1049.396853] env[61995]: value = "task-795183" [ 1049.396853] env[61995]: _type = "Task" [ 1049.396853] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.404760] env[61995]: DEBUG oslo_vmware.api [None req-2b5a7ffd-44bd-4844-81b2-4534c3732094 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795183, 'name': SuspendVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.462566] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.462849] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.463085] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.463284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.463462] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.467877] env[61995]: INFO nova.compute.manager [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Terminating instance [ 1049.469789] env[61995]: DEBUG nova.compute.manager [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.469982] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1049.470838] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc44d7f-80f3-42e6-b8be-7f66e42e8425 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.478637] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1049.478875] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca3ea98b-c39e-47e8-a4a9-11c716d459b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.485155] env[61995]: DEBUG oslo_vmware.api [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1049.485155] env[61995]: value = "task-795184" [ 1049.485155] env[61995]: _type = "Task" [ 1049.485155] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.492637] env[61995]: DEBUG oslo_vmware.api [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.612524] env[61995]: DEBUG nova.scheduler.client.report [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.634457] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f4f0d8-7ff6-a818-4de3-d1296a1d320c, 'name': SearchDatastore_Task, 'duration_secs': 0.007271} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.636022] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b74a8b5-f93f-4a9c-a72c-53cdf7708b3c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.642558] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1049.642558] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52746d5f-dd37-d2a4-ee9d-dc48b5a75f3f" [ 1049.642558] env[61995]: _type = "Task" [ 1049.642558] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.651307] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52746d5f-dd37-d2a4-ee9d-dc48b5a75f3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.907191] env[61995]: DEBUG oslo_vmware.api [None req-2b5a7ffd-44bd-4844-81b2-4534c3732094 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795183, 'name': SuspendVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.995698] env[61995]: DEBUG oslo_vmware.api [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795184, 'name': PowerOffVM_Task, 'duration_secs': 0.211348} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.996037] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1049.996167] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1049.996425] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c3a32d2-1fb3-498b-a160-20e3b493b8d1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.069350] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1050.069602] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1050.069761] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleting the datastore file [datastore2] 432f4d05-ea4d-402c-8934-3aa1aaf93d48 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.070042] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d6882b0-cab9-45c9-a0f5-afc415186124 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.076645] env[61995]: DEBUG oslo_vmware.api [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1050.076645] env[61995]: value = "task-795186" [ 1050.076645] env[61995]: _type = "Task" [ 1050.076645] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.083746] env[61995]: DEBUG oslo_vmware.api [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795186, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.118986] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.121352] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.004s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.122906] env[61995]: INFO nova.compute.claims [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.147233] env[61995]: INFO nova.scheduler.client.report [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Deleted allocations for instance e835ac53-29af-4bd0-b186-5c6270ccf760 [ 1050.163453] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52746d5f-dd37-d2a4-ee9d-dc48b5a75f3f, 'name': SearchDatastore_Task, 'duration_secs': 0.018109} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.164381] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.164960] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c3b2e000-cfa8-4b5b-8763-982c6b768621/c3b2e000-cfa8-4b5b-8763-982c6b768621.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1050.165292] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81b3a570-c3ed-4936-b374-e8955e706273 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.173660] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1050.173660] env[61995]: value = "task-795187" [ 1050.173660] env[61995]: _type = "Task" [ 1050.173660] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.182214] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.409659] env[61995]: DEBUG oslo_vmware.api [None req-2b5a7ffd-44bd-4844-81b2-4534c3732094 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795183, 'name': SuspendVM_Task, 'duration_secs': 0.866302} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.410013] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5a7ffd-44bd-4844-81b2-4534c3732094 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Suspended the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1050.410171] env[61995]: DEBUG nova.compute.manager [None req-2b5a7ffd-44bd-4844-81b2-4534c3732094 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.410984] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29237434-21d8-4a4a-8cd5-4c683bada657 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.590441] env[61995]: DEBUG oslo_vmware.api [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795186, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158809} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.590878] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.591218] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1050.591510] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1050.591801] env[61995]: INFO nova.compute.manager [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1050.592140] env[61995]: DEBUG oslo.service.loopingcall [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.592488] env[61995]: DEBUG nova.compute.manager [-] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.592638] env[61995]: DEBUG nova.network.neutron [-] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1050.661279] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8acec315-5b46-4de5-a437-c64e7b008cdd tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "e835ac53-29af-4bd0-b186-5c6270ccf760" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.742s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.684564] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795187, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.022901] env[61995]: DEBUG nova.compute.manager [req-bb6e5121-1bc2-4e03-85db-af84ff2e67aa req-5ab3a34d-04c3-4c75-acca-48775ff17bb5 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Received event network-vif-deleted-f5f8ea24-7208-4f4c-bec1-77e13198d09c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1051.023247] env[61995]: INFO nova.compute.manager [req-bb6e5121-1bc2-4e03-85db-af84ff2e67aa req-5ab3a34d-04c3-4c75-acca-48775ff17bb5 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Neutron deleted interface f5f8ea24-7208-4f4c-bec1-77e13198d09c; detaching it from the instance and deleting it from the info cache [ 1051.023408] env[61995]: DEBUG nova.network.neutron [req-bb6e5121-1bc2-4e03-85db-af84ff2e67aa req-5ab3a34d-04c3-4c75-acca-48775ff17bb5 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.032913] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.033169] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.185722] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795187, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544923} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.186335] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] c3b2e000-cfa8-4b5b-8763-982c6b768621/c3b2e000-cfa8-4b5b-8763-982c6b768621.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1051.186876] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1051.186976] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af6d3e19-9935-47c9-87f9-1987c96e17af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.195123] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1051.195123] env[61995]: value = "task-795188" [ 1051.195123] env[61995]: _type = "Task" [ 1051.195123] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.209489] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795188, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.379206] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda48352-98f9-4e7f-9598-39155928d0d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.389203] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf304589-1261-43ad-ba02-a8f68e3cec86 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.419584] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4b9522-0b0f-4527-a2d6-98752f7a596f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.426957] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ce789f-7f3c-40f7-be57-b720c9426404 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.442207] env[61995]: DEBUG nova.compute.provider_tree [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.490801] env[61995]: DEBUG nova.network.neutron [-] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.528558] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2f4aca7-e21e-476f-9228-436fbb5b2d57 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.535510] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ae748f-f61c-4b9f-81d9-ed3dcc870b50 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.547344] env[61995]: DEBUG nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.570876] env[61995]: DEBUG nova.compute.manager [req-bb6e5121-1bc2-4e03-85db-af84ff2e67aa req-5ab3a34d-04c3-4c75-acca-48775ff17bb5 service nova] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Detach interface failed, port_id=f5f8ea24-7208-4f4c-bec1-77e13198d09c, reason: Instance 432f4d05-ea4d-402c-8934-3aa1aaf93d48 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1051.709716] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795188, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11305} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.710344] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.711769] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af656385-0b9e-4c2e-a68f-032a5cc65922 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.733428] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] c3b2e000-cfa8-4b5b-8763-982c6b768621/c3b2e000-cfa8-4b5b-8763-982c6b768621.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.734179] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecc63490-5dee-4738-90c0-89e08b437f91 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.753339] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1051.753339] env[61995]: value = "task-795189" [ 1051.753339] env[61995]: _type = "Task" [ 1051.753339] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.761967] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795189, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.946206] env[61995]: DEBUG nova.scheduler.client.report [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.993290] env[61995]: INFO nova.compute.manager [-] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Took 1.40 seconds to deallocate network for instance. [ 1052.040834] env[61995]: DEBUG oslo_concurrency.lockutils [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.041201] env[61995]: DEBUG oslo_concurrency.lockutils [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.067727] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.264794] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795189, 'name': ReconfigVM_Task, 'duration_secs': 0.29954} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.265255] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Reconfigured VM instance instance-00000060 to attach disk [datastore1] c3b2e000-cfa8-4b5b-8763-982c6b768621/c3b2e000-cfa8-4b5b-8763-982c6b768621.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.266027] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbe8f818-75c2-42ca-bc03-7fe6a115b5d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.272847] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1052.272847] env[61995]: value = "task-795190" [ 1052.272847] env[61995]: _type = "Task" [ 1052.272847] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.281541] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795190, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.452252] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.452827] env[61995]: DEBUG nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1052.455581] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.935s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.455808] env[61995]: DEBUG nova.objects.instance [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'resources' on Instance uuid e8bde834-8f55-48d6-8623-1c6e43072c23 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.502040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.545643] env[61995]: DEBUG nova.compute.utils [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1052.669812] env[61995]: DEBUG nova.compute.manager [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.670576] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5eb5a2-3f2e-4209-a9f9-86a82f558fd2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.773694] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.773932] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.784605] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795190, 'name': Rename_Task, 'duration_secs': 0.140098} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.785458] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1052.785713] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d41c3a1-c2a6-4a70-9ec5-5031ce12aa45 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.792467] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1052.792467] env[61995]: value = "task-795191" [ 1052.792467] env[61995]: _type = "Task" [ 1052.792467] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.800647] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.961844] env[61995]: DEBUG nova.compute.utils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1052.963580] env[61995]: DEBUG nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1052.963749] env[61995]: DEBUG nova.network.neutron [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1053.005613] env[61995]: DEBUG nova.policy [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f27aa509c254e2fa6a5d833b53f50e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4042c97b25a24bf0a1ebe23e958f95ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1053.048353] env[61995]: DEBUG oslo_concurrency.lockutils [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.181748] env[61995]: INFO nova.compute.manager [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] instance snapshotting [ 1053.181748] env[61995]: WARNING nova.compute.manager [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 1053.184859] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48eac90a-1acf-4ace-aa1f-3101d9054c71 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.189209] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c495ca-f4b8-43ee-9694-24615cc34031 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.213047] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eb0bb1-e75a-4c8a-b471-e6f5ecd4275c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.216892] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2dfb95-c242-4b73-957f-bdfac6590450 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.250867] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2904aa01-adfe-40ff-a119-11330f5e645c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.258613] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df556df9-a893-4478-bc43-5e817578ec46 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.272504] env[61995]: DEBUG nova.compute.provider_tree [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.275968] env[61995]: DEBUG nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1053.303061] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795191, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.392878] env[61995]: DEBUG nova.network.neutron [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Successfully created port: 534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.467276] env[61995]: DEBUG nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1053.752017] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1053.752017] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-04c1d585-53b7-4e26-a8da-961dc686617f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.759731] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1053.759731] env[61995]: value = "task-795192" [ 1053.759731] env[61995]: _type = "Task" [ 1053.759731] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.769087] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795192, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.775995] env[61995]: DEBUG nova.scheduler.client.report [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.798673] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.804062] env[61995]: DEBUG oslo_vmware.api [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795191, 'name': PowerOnVM_Task, 'duration_secs': 0.548594} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.804358] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1053.804611] env[61995]: INFO nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Took 9.11 seconds to spawn the instance on the hypervisor. [ 1053.804832] env[61995]: DEBUG nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.805641] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49b6f0c-db3d-441a-95cc-401986bdd150 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.119237] env[61995]: DEBUG oslo_concurrency.lockutils [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.119517] env[61995]: DEBUG oslo_concurrency.lockutils [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.119764] env[61995]: INFO nova.compute.manager [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Attaching volume 77915330-bd03-4a88-892c-dc7171c32e4a to /dev/sdb [ 1054.152106] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d400cb14-f760-407c-b6e2-1cd3cd5c8812 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.159369] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c396dd58-5226-4ec0-988f-c7f654b4811a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.172871] env[61995]: DEBUG nova.virt.block_device [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating existing volume attachment record: a411fa2b-305d-4375-9947-13de6bd25dec {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1054.270746] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795192, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.281696] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.283972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.419s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.305823] env[61995]: INFO nova.scheduler.client.report [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted allocations for instance e8bde834-8f55-48d6-8623-1c6e43072c23 [ 1054.323981] env[61995]: INFO nova.compute.manager [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Took 16.49 seconds to build instance. [ 1054.477870] env[61995]: DEBUG nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1054.506430] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.506756] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.506962] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.507197] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.507369] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.507532] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.507766] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.507944] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.508269] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.508481] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.508690] env[61995]: DEBUG nova.virt.hardware [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.509650] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbea614-40f0-4e3b-aa1e-1a0ad5dc4ca2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.517782] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e102f07e-ed6c-4128-b384-807209a75b43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.772700] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795192, 'name': CreateSnapshot_Task, 'duration_secs': 0.647153} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.773155] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1054.774090] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8b9451-0305-4232-b700-b223c5285f57 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.820509] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f66d9f3-26dc-4444-8f91-4baf5408058e tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "e8bde834-8f55-48d6-8623-1c6e43072c23" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.358s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.828396] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5d2d5203-4bd3-4706-b404-46dcfd8d2001 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.006s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.970180] env[61995]: DEBUG nova.network.neutron [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Successfully updated port: 534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1055.009669] env[61995]: DEBUG nova.compute.manager [req-0d39dccf-c8b7-41a6-b1db-de578eae8003 req-a7c214c5-c18a-44f6-a27c-f70a3d38048a service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Received event network-vif-plugged-534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.009669] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d39dccf-c8b7-41a6-b1db-de578eae8003 req-a7c214c5-c18a-44f6-a27c-f70a3d38048a service nova] Acquiring lock "e2014437-fbcd-454a-893f-9ad7c7d461d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.009669] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d39dccf-c8b7-41a6-b1db-de578eae8003 req-a7c214c5-c18a-44f6-a27c-f70a3d38048a service nova] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.009669] env[61995]: DEBUG oslo_concurrency.lockutils [req-0d39dccf-c8b7-41a6-b1db-de578eae8003 req-a7c214c5-c18a-44f6-a27c-f70a3d38048a service nova] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.009669] env[61995]: DEBUG nova.compute.manager [req-0d39dccf-c8b7-41a6-b1db-de578eae8003 req-a7c214c5-c18a-44f6-a27c-f70a3d38048a service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] No waiting events found dispatching network-vif-plugged-534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.009669] env[61995]: WARNING nova.compute.manager [req-0d39dccf-c8b7-41a6-b1db-de578eae8003 req-a7c214c5-c18a-44f6-a27c-f70a3d38048a service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Received unexpected event network-vif-plugged-534906ba-0f5d-4b75-91c1-ac34f5afd2ff for instance with vm_state building and task_state spawning. [ 1055.029393] env[61995]: DEBUG nova.compute.manager [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.029585] env[61995]: DEBUG nova.compute.manager [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing instance network info cache due to event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1055.031085] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.031085] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.031085] env[61995]: DEBUG nova.network.neutron [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1055.041504] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89765d94-5fba-4fbc-ba36-003f4f78dbc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.052330] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3414e56c-fc31-4d0d-9b8e-f2d2256186fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.084608] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14d70c6-134f-4030-a4ec-801501af9aac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.092215] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb93856-71bf-4c68-86f6-c13cfe23b3b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.108683] env[61995]: DEBUG nova.compute.provider_tree [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.292775] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1055.293084] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9354f334-d0aa-4fd4-9cba-8eff5fd684e6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.301832] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1055.301832] env[61995]: value = "task-795196" [ 1055.301832] env[61995]: _type = "Task" [ 1055.301832] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.309873] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795196, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.472838] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.473031] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.473219] env[61995]: DEBUG nova.network.neutron [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1055.613391] env[61995]: DEBUG nova.scheduler.client.report [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.617041] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.620024] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.794225] env[61995]: DEBUG nova.network.neutron [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updated VIF entry in instance network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1055.794710] env[61995]: DEBUG nova.network.neutron [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.813428] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795196, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.007103] env[61995]: DEBUG nova.network.neutron [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1056.121073] env[61995]: DEBUG nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.135261] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.135499] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.164340] env[61995]: DEBUG nova.network.neutron [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Updating instance_info_cache with network_info: [{"id": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "address": "fa:16:3e:99:a8:b7", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap534906ba-0f", "ovs_interfaceid": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.297281] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f610119-6265-45fb-9e09-18bb22e1f3d9 req-5aa635be-a3f6-41cc-8b78-c749ef92ea70 service nova] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.312422] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795196, 'name': CloneVM_Task} progress is 95%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.631415] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.347s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.637983] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.570s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.639487] env[61995]: INFO nova.compute.claims [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.642067] env[61995]: DEBUG nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.658879] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.667535] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.667841] env[61995]: DEBUG nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Instance network_info: |[{"id": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "address": "fa:16:3e:99:a8:b7", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap534906ba-0f", "ovs_interfaceid": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1056.668350] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:a8:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3291573-fad8-48cc-a965-c3554e7cee4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '534906ba-0f5d-4b75-91c1-ac34f5afd2ff', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.676581] env[61995]: DEBUG oslo.service.loopingcall [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.676812] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1056.677834] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab590bdb-69ad-4b3b-a75a-f26133a7feae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.699927] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.699927] env[61995]: value = "task-795198" [ 1056.699927] env[61995]: _type = "Task" [ 1056.699927] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.708029] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795198, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.814881] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795196, 'name': CloneVM_Task, 'duration_secs': 1.283794} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.814881] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Created linked-clone VM from snapshot [ 1056.815147] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6eca51-c87c-47cc-9999-8ebfcab5f23f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.822347] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Uploading image f52129c4-0d21-41df-aad0-db5e438b6b75 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1056.848197] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1056.848197] env[61995]: value = "vm-185409" [ 1056.848197] env[61995]: _type = "VirtualMachine" [ 1056.848197] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1056.848544] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-434795d0-4af2-41ee-92b4-d9093f4d4c02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.855814] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease: (returnval){ [ 1056.855814] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dd2c22-9713-8ed5-3d26-98073b7bf8bc" [ 1056.855814] env[61995]: _type = "HttpNfcLease" [ 1056.855814] env[61995]: } obtained for exporting VM: (result){ [ 1056.855814] env[61995]: value = "vm-185409" [ 1056.855814] env[61995]: _type = "VirtualMachine" [ 1056.855814] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1056.856128] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the lease: (returnval){ [ 1056.856128] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dd2c22-9713-8ed5-3d26-98073b7bf8bc" [ 1056.856128] env[61995]: _type = "HttpNfcLease" [ 1056.856128] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1056.862450] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1056.862450] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dd2c22-9713-8ed5-3d26-98073b7bf8bc" [ 1056.862450] env[61995]: _type = "HttpNfcLease" [ 1056.862450] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1057.034549] env[61995]: DEBUG nova.compute.manager [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Received event network-changed-534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.034838] env[61995]: DEBUG nova.compute.manager [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Refreshing instance network info cache due to event network-changed-534906ba-0f5d-4b75-91c1-ac34f5afd2ff. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1057.035114] env[61995]: DEBUG oslo_concurrency.lockutils [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] Acquiring lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.035314] env[61995]: DEBUG oslo_concurrency.lockutils [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] Acquired lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.035541] env[61995]: DEBUG nova.network.neutron [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Refreshing network info cache for port 534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.168424] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.204296] env[61995]: INFO nova.scheduler.client.report [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted allocation for migration 5210f5ef-264f-4b03-9de0-5d58b3b52ac2 [ 1057.211606] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795198, 'name': CreateVM_Task, 'duration_secs': 0.36599} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.211956] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1057.212619] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.212787] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.213119] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.213366] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bf4f054-6530-4bd0-8cf0-bb28e4c44f3a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.218283] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1057.218283] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52caa390-0eef-9206-82aa-c25ef9c758f3" [ 1057.218283] env[61995]: _type = "Task" [ 1057.218283] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.226153] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52caa390-0eef-9206-82aa-c25ef9c758f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.364971] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1057.364971] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dd2c22-9713-8ed5-3d26-98073b7bf8bc" [ 1057.364971] env[61995]: _type = "HttpNfcLease" [ 1057.364971] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1057.364971] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1057.364971] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dd2c22-9713-8ed5-3d26-98073b7bf8bc" [ 1057.364971] env[61995]: _type = "HttpNfcLease" [ 1057.364971] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1057.365626] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40950a73-cc89-4f86-a0cc-aec7e759a57f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.373273] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ea688a-f7d9-00c7-852f-7ff640718bba/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1057.373459] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ea688a-f7d9-00c7-852f-7ff640718bba/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1057.465694] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4d1be8d0-c6f7-4b67-8e05-fc9e7711fc19 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.713191] env[61995]: DEBUG oslo_concurrency.lockutils [None req-35b21d85-40cf-4706-9654-90c99b03477a tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 17.404s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.736045] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52caa390-0eef-9206-82aa-c25ef9c758f3, 'name': SearchDatastore_Task, 'duration_secs': 0.018102} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.736612] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.736698] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.737094] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.737302] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.737901] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.737982] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9285af8f-5e15-475a-b15e-330969fe3414 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.747823] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.747823] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1057.751057] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7454908f-aea3-4130-abf1-b6a11110bb8d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.755375] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1057.755375] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e5dd30-e861-819d-e4de-e53793976fd1" [ 1057.755375] env[61995]: _type = "Task" [ 1057.755375] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.771599] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e5dd30-e861-819d-e4de-e53793976fd1, 'name': SearchDatastore_Task, 'duration_secs': 0.01123} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.772443] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdfc98cd-688e-4674-8419-2b4a117f6430 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.777671] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1057.777671] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c11764-f3b9-20db-44c6-9b2cec5d9fa8" [ 1057.777671] env[61995]: _type = "Task" [ 1057.777671] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.789192] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c11764-f3b9-20db-44c6-9b2cec5d9fa8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.799964] env[61995]: DEBUG nova.network.neutron [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Updated VIF entry in instance network info cache for port 534906ba-0f5d-4b75-91c1-ac34f5afd2ff. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1057.800367] env[61995]: DEBUG nova.network.neutron [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Updating instance_info_cache with network_info: [{"id": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "address": "fa:16:3e:99:a8:b7", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap534906ba-0f", "ovs_interfaceid": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.905283] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5365b2ff-a3ae-4001-89b5-8745045e95dd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.913347] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d24fafe-d2b0-41db-b3a1-e1b2c14a5e11 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.945549] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726d8204-46fd-44d6-aa27-f2f056783d4d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.953858] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1181eef6-14f7-4228-8e5e-cd5e59f01e9a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.970088] env[61995]: DEBUG nova.compute.provider_tree [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.026687] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.026973] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.027213] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "d9c53806-68c0-4872-a262-36bc05573674-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.027539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.027874] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.030345] env[61995]: INFO nova.compute.manager [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Terminating instance [ 1058.032227] env[61995]: DEBUG nova.compute.manager [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1058.032423] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1058.033359] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc13e079-3295-49e1-928f-1cb2832ce512 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.041144] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1058.041378] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c6e43f9-d6e0-4b1b-951b-a428574b813f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.048119] env[61995]: DEBUG oslo_vmware.api [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1058.048119] env[61995]: value = "task-795200" [ 1058.048119] env[61995]: _type = "Task" [ 1058.048119] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.056385] env[61995]: DEBUG oslo_vmware.api [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.289956] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c11764-f3b9-20db-44c6-9b2cec5d9fa8, 'name': SearchDatastore_Task, 'duration_secs': 0.010163} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.290295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.290621] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] e2014437-fbcd-454a-893f-9ad7c7d461d0/e2014437-fbcd-454a-893f-9ad7c7d461d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1058.290933] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55122468-1df7-4cc1-859e-8c1e1abffd2d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.298019] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1058.298019] env[61995]: value = "task-795201" [ 1058.298019] env[61995]: _type = "Task" [ 1058.298019] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.303412] env[61995]: DEBUG oslo_concurrency.lockutils [req-fda1dec7-4d53-40c7-afca-55849bf60a77 req-00188272-e7b0-4eb5-adfa-336ac17b5e6c service nova] Releasing lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.306922] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795201, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.475677] env[61995]: DEBUG nova.scheduler.client.report [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.560490] env[61995]: DEBUG oslo_vmware.api [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795200, 'name': PowerOffVM_Task, 'duration_secs': 0.234429} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.560855] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1058.561259] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1058.561577] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5ed5c5e-4695-47ca-bc25-0f41dc39d8c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.655440] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1058.655709] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1058.655869] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleting the datastore file [datastore2] d9c53806-68c0-4872-a262-36bc05573674 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.656187] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05896947-6af5-4cc4-a5ee-6413217ed3c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.665154] env[61995]: DEBUG oslo_vmware.api [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1058.665154] env[61995]: value = "task-795203" [ 1058.665154] env[61995]: _type = "Task" [ 1058.665154] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.674811] env[61995]: DEBUG oslo_vmware.api [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.722568] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1058.722568] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185408', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'name': 'volume-77915330-bd03-4a88-892c-dc7171c32e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66c56f7b-1fc7-4aed-8afc-350817e1ca48', 'attached_at': '', 'detached_at': '', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'serial': '77915330-bd03-4a88-892c-dc7171c32e4a'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1058.722955] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da8562d-78b0-47ee-a8eb-bdde230a3352 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.743083] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00601b7d-3a6e-4e1e-b1a1-334a8acb9609 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.773059] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] volume-77915330-bd03-4a88-892c-dc7171c32e4a/volume-77915330-bd03-4a88-892c-dc7171c32e4a.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1058.773484] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59a1d7a7-bec8-41e1-9856-5e00ed03a9ba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.793693] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1058.793693] env[61995]: value = "task-795204" [ 1058.793693] env[61995]: _type = "Task" [ 1058.793693] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.807315] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795204, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.812468] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795201, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.983159] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.983770] env[61995]: DEBUG nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1058.987907] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.486s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.988491] env[61995]: DEBUG nova.objects.instance [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lazy-loading 'resources' on Instance uuid 432f4d05-ea4d-402c-8934-3aa1aaf93d48 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.175993] env[61995]: DEBUG oslo_vmware.api [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361869} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.176309] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.176521] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1059.176728] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1059.176913] env[61995]: INFO nova.compute.manager [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: d9c53806-68c0-4872-a262-36bc05573674] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1059.177187] env[61995]: DEBUG oslo.service.loopingcall [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.177387] env[61995]: DEBUG nova.compute.manager [-] [instance: d9c53806-68c0-4872-a262-36bc05573674] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1059.177480] env[61995]: DEBUG nova.network.neutron [-] [instance: d9c53806-68c0-4872-a262-36bc05573674] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1059.307983] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795204, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.311230] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795201, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517656} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.311484] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] e2014437-fbcd-454a-893f-9ad7c7d461d0/e2014437-fbcd-454a-893f-9ad7c7d461d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1059.311701] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1059.311949] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff980c29-23c1-4896-952d-e485c2a843b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.319243] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1059.319243] env[61995]: value = "task-795205" [ 1059.319243] env[61995]: _type = "Task" [ 1059.319243] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.327030] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795205, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.492302] env[61995]: DEBUG nova.compute.utils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.496902] env[61995]: DEBUG nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1059.497144] env[61995]: DEBUG nova.network.neutron [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1059.541060] env[61995]: DEBUG nova.policy [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3efdbd0f8e0340dc8e4bfbf672d62f0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83064cbae17c429d8d084837635486da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1059.748808] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af9444e-a469-4064-83a3-a977263ac8bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.757041] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19d5cfe-3ae5-4974-a6fb-43fece057856 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.787915] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e25ced-b5b3-4c7c-ab10-7925626b84c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.796545] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b3a0e1-3bf7-4d35-bfc6-4579d4c93d69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.809809] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795204, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.818367] env[61995]: DEBUG nova.compute.provider_tree [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.827772] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795205, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08097} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.828678] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1059.829459] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfba4be-6b91-4a97-be9d-2bed1f59d17b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.854287] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] e2014437-fbcd-454a-893f-9ad7c7d461d0/e2014437-fbcd-454a-893f-9ad7c7d461d0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.854287] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e371380f-cbb8-40bb-bc90-525b44a8a3c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.874352] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1059.874352] env[61995]: value = "task-795206" [ 1059.874352] env[61995]: _type = "Task" [ 1059.874352] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.885019] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795206, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.885019] env[61995]: DEBUG nova.network.neutron [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Successfully created port: 7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.939985] env[61995]: DEBUG nova.compute.manager [req-446d5d33-ceea-46a0-aa22-29d28b1c1f69 req-187dfce1-d969-4a34-b992-9c2faf0ac333 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Received event network-vif-deleted-802b752a-238e-4b55-8bd3-f73eddc4b0f4 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1059.940260] env[61995]: INFO nova.compute.manager [req-446d5d33-ceea-46a0-aa22-29d28b1c1f69 req-187dfce1-d969-4a34-b992-9c2faf0ac333 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Neutron deleted interface 802b752a-238e-4b55-8bd3-f73eddc4b0f4; detaching it from the instance and deleting it from the info cache [ 1059.940419] env[61995]: DEBUG nova.network.neutron [req-446d5d33-ceea-46a0-aa22-29d28b1c1f69 req-187dfce1-d969-4a34-b992-9c2faf0ac333 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.998204] env[61995]: DEBUG nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.315631] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795204, 'name': ReconfigVM_Task, 'duration_secs': 1.45081} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.316301] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfigured VM instance instance-00000058 to attach disk [datastore2] volume-77915330-bd03-4a88-892c-dc7171c32e4a/volume-77915330-bd03-4a88-892c-dc7171c32e4a.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.325291] env[61995]: DEBUG nova.scheduler.client.report [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.330275] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e14affa0-668c-4e97-9a6b-97f6016d78b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.350079] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.362s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.352859] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.554s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.355354] env[61995]: INFO nova.compute.claims [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.367204] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1060.367204] env[61995]: value = "task-795207" [ 1060.367204] env[61995]: _type = "Task" [ 1060.367204] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.382351] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795207, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.384719] env[61995]: INFO nova.scheduler.client.report [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleted allocations for instance 432f4d05-ea4d-402c-8934-3aa1aaf93d48 [ 1060.388696] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795206, 'name': ReconfigVM_Task, 'duration_secs': 0.389154} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.391801] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Reconfigured VM instance instance-00000061 to attach disk [datastore2] e2014437-fbcd-454a-893f-9ad7c7d461d0/e2014437-fbcd-454a-893f-9ad7c7d461d0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.392390] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-582a9fdb-adb5-4e52-8b20-b1cb14befebd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.401214] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1060.401214] env[61995]: value = "task-795208" [ 1060.401214] env[61995]: _type = "Task" [ 1060.401214] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.412022] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795208, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.416597] env[61995]: DEBUG nova.network.neutron [-] [instance: d9c53806-68c0-4872-a262-36bc05573674] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.445272] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f0bc9f15-93e7-41e1-942e-adb008946a41 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.455342] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdec571d-06d4-4054-8abd-9e1a76a07292 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.487287] env[61995]: DEBUG nova.compute.manager [req-446d5d33-ceea-46a0-aa22-29d28b1c1f69 req-187dfce1-d969-4a34-b992-9c2faf0ac333 service nova] [instance: d9c53806-68c0-4872-a262-36bc05573674] Detach interface failed, port_id=802b752a-238e-4b55-8bd3-f73eddc4b0f4, reason: Instance d9c53806-68c0-4872-a262-36bc05573674 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1060.504157] env[61995]: INFO nova.virt.block_device [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Booting with volume bbdbeaec-455e-4027-8c73-ed438ff9dde7 at /dev/sda [ 1060.546873] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e05e45c3-37f1-4714-a8e2-03ac3c8a9b8d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.556255] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2056d196-36c0-4fc3-8bdd-d2e18af3438d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.590874] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d32ecf0-ed5d-4b9a-bcc4-bd90b5e9117d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.599126] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa56719-208b-4088-b26a-e5cde64ec2f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.631401] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0b0e83-ede3-4617-8d50-4d6b4bc2f4b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.638021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c7057f-a4b6-40af-9320-1d5c30c1593c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.651253] env[61995]: DEBUG nova.virt.block_device [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating existing volume attachment record: c88cf9d4-9142-47b6-a301-3de841f75b6f {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1060.876919] env[61995]: DEBUG oslo_vmware.api [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795207, 'name': ReconfigVM_Task, 'duration_secs': 0.168213} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.877325] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185408', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'name': 'volume-77915330-bd03-4a88-892c-dc7171c32e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66c56f7b-1fc7-4aed-8afc-350817e1ca48', 'attached_at': '', 'detached_at': '', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'serial': '77915330-bd03-4a88-892c-dc7171c32e4a'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1060.897328] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b3621bbe-c43c-4484-abf4-155c2b4ccc04 tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "432f4d05-ea4d-402c-8934-3aa1aaf93d48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.434s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.911760] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795208, 'name': Rename_Task, 'duration_secs': 0.160138} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.912740] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1060.913015] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3802f97c-fd29-4101-83c3-bb1d01c16711 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.919448] env[61995]: INFO nova.compute.manager [-] [instance: d9c53806-68c0-4872-a262-36bc05573674] Took 1.74 seconds to deallocate network for instance. [ 1060.923581] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1060.923581] env[61995]: value = "task-795209" [ 1060.923581] env[61995]: _type = "Task" [ 1060.923581] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.938686] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795209, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.433854] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.439667] env[61995]: DEBUG oslo_vmware.api [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795209, 'name': PowerOnVM_Task, 'duration_secs': 0.509687} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.439841] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1061.440545] env[61995]: INFO nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Took 6.96 seconds to spawn the instance on the hypervisor. [ 1061.440545] env[61995]: DEBUG nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1061.441589] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ec9dfa-7ee5-47ff-9c10-3505e46c0d27 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.461287] env[61995]: DEBUG nova.compute.manager [req-6a63b34c-5ae4-4d90-bcfb-8e1eccb9a16b req-68d5dc2b-c717-40fa-ad8d-7073ff411081 service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Received event network-vif-plugged-7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.461287] env[61995]: DEBUG oslo_concurrency.lockutils [req-6a63b34c-5ae4-4d90-bcfb-8e1eccb9a16b req-68d5dc2b-c717-40fa-ad8d-7073ff411081 service nova] Acquiring lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.461287] env[61995]: DEBUG oslo_concurrency.lockutils [req-6a63b34c-5ae4-4d90-bcfb-8e1eccb9a16b req-68d5dc2b-c717-40fa-ad8d-7073ff411081 service nova] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.461287] env[61995]: DEBUG oslo_concurrency.lockutils [req-6a63b34c-5ae4-4d90-bcfb-8e1eccb9a16b req-68d5dc2b-c717-40fa-ad8d-7073ff411081 service nova] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.461287] env[61995]: DEBUG nova.compute.manager [req-6a63b34c-5ae4-4d90-bcfb-8e1eccb9a16b req-68d5dc2b-c717-40fa-ad8d-7073ff411081 service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] No waiting events found dispatching network-vif-plugged-7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.461287] env[61995]: WARNING nova.compute.manager [req-6a63b34c-5ae4-4d90-bcfb-8e1eccb9a16b req-68d5dc2b-c717-40fa-ad8d-7073ff411081 service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Received unexpected event network-vif-plugged-7c06dfad-348a-49b5-8d8d-d606f4b559c9 for instance with vm_state building and task_state block_device_mapping. [ 1061.547663] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "06886222-5f7f-482e-b5ee-afd7326f2c70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.548158] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.548158] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "06886222-5f7f-482e-b5ee-afd7326f2c70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.550138] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.550277] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.554593] env[61995]: INFO nova.compute.manager [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Terminating instance [ 1061.556745] env[61995]: DEBUG nova.compute.manager [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1061.556839] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1061.557773] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62a62f2-9021-4c0c-af12-0183a58e7d21 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.569030] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1061.569030] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6ebde47-b0db-4dbb-ae58-975da784d76f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.570393] env[61995]: DEBUG nova.network.neutron [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Successfully updated port: 7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.577334] env[61995]: DEBUG oslo_vmware.api [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1061.577334] env[61995]: value = "task-795210" [ 1061.577334] env[61995]: _type = "Task" [ 1061.577334] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.590138] env[61995]: DEBUG oslo_vmware.api [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.630931] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c194a1f7-3030-43aa-9dc0-591c1bf69a8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.639158] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad6490e-0c5f-4a0f-91c6-c571253edd1a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.669194] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e42d2d-8dd2-4cdb-a6da-b0b844afcc12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.677221] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c8ebd9-aea8-4621-9b91-d03db8a68771 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.691499] env[61995]: DEBUG nova.compute.provider_tree [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.927411] env[61995]: DEBUG nova.objects.instance [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lazy-loading 'flavor' on Instance uuid 66c56f7b-1fc7-4aed-8afc-350817e1ca48 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.965202] env[61995]: INFO nova.compute.manager [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Took 16.86 seconds to build instance. [ 1062.073073] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.073198] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.073344] env[61995]: DEBUG nova.network.neutron [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1062.088669] env[61995]: DEBUG oslo_vmware.api [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795210, 'name': PowerOffVM_Task, 'duration_secs': 0.207934} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.089331] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1062.089331] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1062.089589] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8068a17c-4e4a-4ed7-8437-052fda9bf446 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.160416] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1062.160674] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1062.160869] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleting the datastore file [datastore2] 06886222-5f7f-482e-b5ee-afd7326f2c70 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1062.161167] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71630488-63e9-4b32-ba77-c9cc31dfe23b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.168726] env[61995]: DEBUG oslo_vmware.api [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for the task: (returnval){ [ 1062.168726] env[61995]: value = "task-795212" [ 1062.168726] env[61995]: _type = "Task" [ 1062.168726] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.176848] env[61995]: DEBUG oslo_vmware.api [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795212, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.195101] env[61995]: DEBUG nova.scheduler.client.report [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.433595] env[61995]: DEBUG oslo_concurrency.lockutils [None req-052a1e1d-483e-4c49-88e4-44bbf5d591dc tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.314s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.467398] env[61995]: DEBUG oslo_concurrency.lockutils [None req-eeaff376-92a6-42a8-a80d-a906b51a70e4 tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.373s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.573301] env[61995]: INFO nova.compute.manager [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Rescuing [ 1062.573301] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.573301] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.573301] env[61995]: DEBUG nova.network.neutron [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1062.607103] env[61995]: DEBUG nova.network.neutron [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1062.681581] env[61995]: DEBUG oslo_vmware.api [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Task: {'id': task-795212, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317182} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.681581] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.681581] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1062.681757] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1062.681808] env[61995]: INFO nova.compute.manager [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1062.682282] env[61995]: DEBUG oslo.service.loopingcall [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.682282] env[61995]: DEBUG nova.compute.manager [-] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1062.682386] env[61995]: DEBUG nova.network.neutron [-] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1062.701030] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.701030] env[61995]: DEBUG nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1062.703332] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.045s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.704831] env[61995]: INFO nova.compute.claims [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.737266] env[61995]: DEBUG nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1062.737877] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.738253] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.738462] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.738663] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.738809] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.739286] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.739884] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.740060] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.740251] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.740429] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.740615] env[61995]: DEBUG nova.virt.hardware [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.741639] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfbc30d-98cb-4412-a300-13783ba0f10f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.751750] env[61995]: DEBUG nova.network.neutron [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance_info_cache with network_info: [{"id": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "address": "fa:16:3e:33:74:c0", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c06dfad-34", "ovs_interfaceid": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.753804] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06eb643-679c-4b08-afa7-331bd69b07e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.078587] env[61995]: DEBUG nova.compute.manager [req-0355cf1c-0814-4703-b74b-1e2d7b79d531 req-7a04bb7f-6cac-4379-80ac-c153d2a0edf7 service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Received event network-vif-deleted-14a93316-278c-4140-a1e5-567a797d3a44 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.078587] env[61995]: INFO nova.compute.manager [req-0355cf1c-0814-4703-b74b-1e2d7b79d531 req-7a04bb7f-6cac-4379-80ac-c153d2a0edf7 service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Neutron deleted interface 14a93316-278c-4140-a1e5-567a797d3a44; detaching it from the instance and deleting it from the info cache [ 1063.078891] env[61995]: DEBUG nova.network.neutron [req-0355cf1c-0814-4703-b74b-1e2d7b79d531 req-7a04bb7f-6cac-4379-80ac-c153d2a0edf7 service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.209371] env[61995]: DEBUG nova.compute.utils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.214217] env[61995]: DEBUG nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1063.214217] env[61995]: DEBUG nova.network.neutron [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1063.259057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.259403] env[61995]: DEBUG nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance network_info: |[{"id": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "address": "fa:16:3e:33:74:c0", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c06dfad-34", "ovs_interfaceid": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1063.259709] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:74:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7c06dfad-348a-49b5-8d8d-d606f4b559c9', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.268328] env[61995]: DEBUG oslo.service.loopingcall [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.271068] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1063.271331] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0872674d-3abe-4d00-aff5-2af9e7b466c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.287438] env[61995]: DEBUG nova.policy [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14d596474e4e48a4889db695a088fc72', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4de205843d14c7a91156c9b49ab7657', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1063.294409] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.294409] env[61995]: value = "task-795213" [ 1063.294409] env[61995]: _type = "Task" [ 1063.294409] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.303031] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795213, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.367650] env[61995]: DEBUG nova.network.neutron [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.489250] env[61995]: DEBUG nova.compute.manager [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Received event network-changed-7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.489561] env[61995]: DEBUG nova.compute.manager [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Refreshing instance network info cache due to event network-changed-7c06dfad-348a-49b5-8d8d-d606f4b559c9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1063.489841] env[61995]: DEBUG oslo_concurrency.lockutils [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] Acquiring lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.490023] env[61995]: DEBUG oslo_concurrency.lockutils [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] Acquired lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.490221] env[61995]: DEBUG nova.network.neutron [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Refreshing network info cache for port 7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1063.548859] env[61995]: DEBUG nova.network.neutron [-] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.586030] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea968f65-f233-4307-aed3-f5e0b0c64cf7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.594418] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3259c82b-accd-4b5a-ac8d-f08b2e621be1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.626058] env[61995]: DEBUG nova.compute.manager [req-0355cf1c-0814-4703-b74b-1e2d7b79d531 req-7a04bb7f-6cac-4379-80ac-c153d2a0edf7 service nova] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Detach interface failed, port_id=14a93316-278c-4140-a1e5-567a797d3a44, reason: Instance 06886222-5f7f-482e-b5ee-afd7326f2c70 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1063.707208] env[61995]: DEBUG nova.network.neutron [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Successfully created port: 2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.713898] env[61995]: DEBUG nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1063.805789] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795213, 'name': CreateVM_Task, 'duration_secs': 0.416996} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.808326] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1063.809334] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185404', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'name': 'volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac81c31d-d310-42c7-a6b4-166f3e82cf56', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'serial': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7'}, 'attachment_id': 'c88cf9d4-9142-47b6-a301-3de841f75b6f', 'guest_format': None, 'boot_index': 0, 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61995) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1063.809490] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Root volume attach. Driver type: vmdk {{(pid=61995) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1063.810290] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f80197-6b7d-4eae-849a-fde5a611de11 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.820512] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba8a5f8-e88e-4ee4-aba7-3a2d14e3b435 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.827231] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0339dae4-0320-40cf-91fb-1c130dd1b8af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.835472] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-b11c56b9-99d0-4679-be15-f8789b291ad2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.842226] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1063.842226] env[61995]: value = "task-795214" [ 1063.842226] env[61995]: _type = "Task" [ 1063.842226] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.852390] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.878077] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.998147] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff00abd6-f1dd-4204-bb69-2f011bcca968 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.005775] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2430224-cd87-4be1-913a-751cc4ed7036 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.039655] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73023e4-0c5b-4c6b-aaae-d92ea3985336 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.047921] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4877538b-7461-4959-bacf-6ae8d0eb45cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.052528] env[61995]: INFO nova.compute.manager [-] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Took 1.37 seconds to deallocate network for instance. [ 1064.069020] env[61995]: DEBUG nova.compute.provider_tree [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.274833] env[61995]: DEBUG nova.network.neutron [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updated VIF entry in instance network info cache for port 7c06dfad-348a-49b5-8d8d-d606f4b559c9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1064.275195] env[61995]: DEBUG nova.network.neutron [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance_info_cache with network_info: [{"id": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "address": "fa:16:3e:33:74:c0", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c06dfad-34", "ovs_interfaceid": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.353042] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task} progress is 43%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.424379] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1064.424648] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e84ca286-6fcb-48f2-83d3-b81acea2dbf7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.432753] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1064.432753] env[61995]: value = "task-795215" [ 1064.432753] env[61995]: _type = "Task" [ 1064.432753] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.442054] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795215, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.573115] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.579631] env[61995]: DEBUG nova.scheduler.client.report [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.724893] env[61995]: DEBUG nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1064.749543] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.749823] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.750176] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.750413] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.750571] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.750725] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.750942] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.751121] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.751336] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.751580] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.751787] env[61995]: DEBUG nova.virt.hardware [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.752720] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734366ba-a368-491c-8ff1-27b4dcd8d4c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.762156] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2f806a-209b-41a3-bbbd-ba5eb3cef2b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.779497] env[61995]: DEBUG oslo_concurrency.lockutils [req-08d2d864-60e8-43f2-8d79-8d11acd0ac8a req-3de93dde-204d-47e8-8801-4ba626e74aac service nova] Releasing lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.854768] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task} progress is 56%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.943848] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795215, 'name': PowerOffVM_Task, 'duration_secs': 0.298202} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.944249] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1064.945239] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f188dead-6820-4991-885b-7a6d2b4727ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.968681] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc55deb3-2605-46f2-a558-7424359f96e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.998739] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1064.999091] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b924a2c-bb3c-469e-b546-fc79cdd9fd75 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.006839] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1065.006839] env[61995]: value = "task-795216" [ 1065.006839] env[61995]: _type = "Task" [ 1065.006839] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.015984] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1065.016201] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.016462] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.016647] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.016876] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.017216] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86f516d2-7916-4697-9b28-4a8cacdf59d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.026865] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.027100] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1065.027915] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c0115c1-3565-44d2-95f5-104ede6c729f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.033343] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1065.033343] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f647c4-5783-e0a3-d07d-ecca58770c30" [ 1065.033343] env[61995]: _type = "Task" [ 1065.033343] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.041821] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f647c4-5783-e0a3-d07d-ecca58770c30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.086348] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.087056] env[61995]: DEBUG nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1065.090548] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.922s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.095479] env[61995]: INFO nova.compute.claims [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1065.113555] env[61995]: DEBUG nova.compute.manager [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Received event network-changed-534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1065.113912] env[61995]: DEBUG nova.compute.manager [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Refreshing instance network info cache due to event network-changed-534906ba-0f5d-4b75-91c1-ac34f5afd2ff. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1065.114154] env[61995]: DEBUG oslo_concurrency.lockutils [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] Acquiring lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.114325] env[61995]: DEBUG oslo_concurrency.lockutils [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] Acquired lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.114448] env[61995]: DEBUG nova.network.neutron [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Refreshing network info cache for port 534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1065.297239] env[61995]: DEBUG nova.network.neutron [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Successfully updated port: 2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.353413] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task} progress is 69%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.518838] env[61995]: DEBUG nova.compute.manager [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Received event network-vif-plugged-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1065.519198] env[61995]: DEBUG oslo_concurrency.lockutils [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.519360] env[61995]: DEBUG oslo_concurrency.lockutils [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.519561] env[61995]: DEBUG oslo_concurrency.lockutils [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.519733] env[61995]: DEBUG nova.compute.manager [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] No waiting events found dispatching network-vif-plugged-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1065.519907] env[61995]: WARNING nova.compute.manager [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Received unexpected event network-vif-plugged-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 for instance with vm_state building and task_state spawning. [ 1065.520272] env[61995]: DEBUG nova.compute.manager [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Received event network-changed-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1065.520482] env[61995]: DEBUG nova.compute.manager [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Refreshing instance network info cache due to event network-changed-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1065.520680] env[61995]: DEBUG oslo_concurrency.lockutils [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] Acquiring lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.520820] env[61995]: DEBUG oslo_concurrency.lockutils [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] Acquired lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.520979] env[61995]: DEBUG nova.network.neutron [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Refreshing network info cache for port 2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1065.543602] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f647c4-5783-e0a3-d07d-ecca58770c30, 'name': SearchDatastore_Task, 'duration_secs': 0.01254} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.544416] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07716cb4-aa5d-4773-a771-872acc72367a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.551186] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1065.551186] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b61ceb-5e33-c1d8-4f75-25503dabb6db" [ 1065.551186] env[61995]: _type = "Task" [ 1065.551186] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.560494] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b61ceb-5e33-c1d8-4f75-25503dabb6db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.602501] env[61995]: DEBUG nova.compute.utils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.606700] env[61995]: DEBUG nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.606930] env[61995]: DEBUG nova.network.neutron [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1065.680722] env[61995]: DEBUG nova.policy [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'caf9bbbd848d4d7ba53ba3bd553943e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2018aae65cf4bf98b5ac82519201cf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1065.800368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.853401] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task} progress is 84%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.915108] env[61995]: DEBUG nova.network.neutron [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Updated VIF entry in instance network info cache for port 534906ba-0f5d-4b75-91c1-ac34f5afd2ff. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1065.915534] env[61995]: DEBUG nova.network.neutron [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Updating instance_info_cache with network_info: [{"id": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "address": "fa:16:3e:99:a8:b7", "network": {"id": "5e43abf2-cc28-4f84-b41c-e9be0fe29d5c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1680434625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4042c97b25a24bf0a1ebe23e958f95ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap534906ba-0f", "ovs_interfaceid": "534906ba-0f5d-4b75-91c1-ac34f5afd2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.044650] env[61995]: DEBUG nova.network.neutron [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Successfully created port: 6d70ab82-17c8-4f87-a1da-012146ad2eee {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1066.056681] env[61995]: DEBUG nova.network.neutron [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1066.063242] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b61ceb-5e33-c1d8-4f75-25503dabb6db, 'name': SearchDatastore_Task, 'duration_secs': 0.013209} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.063890] env[61995]: DEBUG oslo_concurrency.lockutils [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.063890] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk. {{(pid=61995) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1066.065088] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1459b13e-73de-4903-a23f-18f2f953d6be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.072617] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1066.072617] env[61995]: value = "task-795217" [ 1066.072617] env[61995]: _type = "Task" [ 1066.072617] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.084302] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.111599] env[61995]: DEBUG nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1066.161834] env[61995]: DEBUG nova.network.neutron [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.358423] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.364345] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff25af0c-ff24-4d24-af97-8a85d98a78d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.372382] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db00499-d480-4cc5-a466-00c5a593a0a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.407941] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32c2d65-69a7-4ae2-9647-5b7bec34ae39 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.416446] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2e1327-74bb-4ab1-8e17-6484e6e0225b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.421086] env[61995]: DEBUG oslo_concurrency.lockutils [req-3efd0deb-0dac-4300-9890-d8919339d603 req-a1f33dde-d5fa-45b1-be75-6c8b6c6fd30c service nova] Releasing lock "refresh_cache-e2014437-fbcd-454a-893f-9ad7c7d461d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.431693] env[61995]: DEBUG nova.compute.provider_tree [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.463217] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ea688a-f7d9-00c7-852f-7ff640718bba/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1066.464283] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f76b15-f284-4a51-95bd-959f5cf6a930 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.472467] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ea688a-f7d9-00c7-852f-7ff640718bba/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1066.472750] env[61995]: ERROR oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ea688a-f7d9-00c7-852f-7ff640718bba/disk-0.vmdk due to incomplete transfer. [ 1066.473032] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c66bac42-6688-4ca1-956d-1f6bc423d3bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.481788] env[61995]: DEBUG oslo_vmware.rw_handles [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ea688a-f7d9-00c7-852f-7ff640718bba/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1066.482100] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Uploaded image f52129c4-0d21-41df-aad0-db5e438b6b75 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1066.484611] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1066.484897] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-01e6af16-bdb8-4bb7-8c2b-5014ecddd046 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.491251] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1066.491251] env[61995]: value = "task-795218" [ 1066.491251] env[61995]: _type = "Task" [ 1066.491251] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.500390] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795218, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.584826] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795217, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.664326] env[61995]: DEBUG oslo_concurrency.lockutils [req-3fdcc4b0-4b80-48dd-9247-a14ffd1e6edb req-b8eea9cd-6db1-4da3-bd96-c8b761b6ccb3 service nova] Releasing lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.664700] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.664862] env[61995]: DEBUG nova.network.neutron [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1066.855058] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task} progress is 98%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.935344] env[61995]: DEBUG nova.scheduler.client.report [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.001923] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795218, 'name': Destroy_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.083597] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541553} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.084053] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk. [ 1067.084696] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77306206-8426-46b3-b254-38e5f3c39009 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.112733] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.113054] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3da6426a-34c0-4386-ba18-da9d68f4f607 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.126698] env[61995]: DEBUG nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1067.135768] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1067.135768] env[61995]: value = "task-795219" [ 1067.135768] env[61995]: _type = "Task" [ 1067.135768] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.145172] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795219, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.152333] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1067.152588] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1067.152768] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1067.152962] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1067.153127] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1067.153282] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1067.153491] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1067.153758] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1067.153985] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1067.154238] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1067.154359] env[61995]: DEBUG nova.virt.hardware [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1067.155777] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a2fcaa-b93a-4bc6-827b-c37591573a3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.162996] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10f8300-efde-4ca0-8ed9-01c245fbaa26 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.187028] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.187156] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.187331] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.187517] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.187689] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.189770] env[61995]: INFO nova.compute.manager [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Terminating instance [ 1067.191538] env[61995]: DEBUG nova.compute.manager [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.191727] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1067.192509] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e931bd-6fc8-4134-ae02-01395dce0133 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.199524] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1067.199745] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4a34ad0-fc1f-48dd-9f43-b50681994fdd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.203253] env[61995]: DEBUG nova.network.neutron [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1067.285926] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1067.286182] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1067.286380] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleting the datastore file [datastore1] ff07d5b8-38b4-48d4-a6d0-632d37d3a91a {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.286663] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-110ef77e-1170-4933-ade8-7e55b2a0d8ce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.292481] env[61995]: DEBUG oslo_vmware.api [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1067.292481] env[61995]: value = "task-795221" [ 1067.292481] env[61995]: _type = "Task" [ 1067.292481] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.299938] env[61995]: DEBUG oslo_vmware.api [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795221, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.349409] env[61995]: DEBUG nova.network.neutron [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updating instance_info_cache with network_info: [{"id": "2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7", "address": "fa:16:3e:dd:33:25", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ebf2f2d-13", "ovs_interfaceid": "2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.355916] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795214, 'name': RelocateVM_Task, 'duration_secs': 3.30119} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.356264] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1067.356477] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185404', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'name': 'volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac81c31d-d310-42c7-a6b4-166f3e82cf56', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'serial': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1067.357286] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ebbf81-6f81-4c49-ad63-dbd164ce820c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.373675] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f292a14-e500-46fd-a232-9ecd985aafd4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.396816] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7/volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.397447] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32b93cea-382c-4d7f-b47e-faf70ade3de2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.421024] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1067.421024] env[61995]: value = "task-795222" [ 1067.421024] env[61995]: _type = "Task" [ 1067.421024] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.428168] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795222, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.440897] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.441525] env[61995]: DEBUG nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1067.444429] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.011s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.444639] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.446834] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.874s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.447077] env[61995]: DEBUG nova.objects.instance [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lazy-loading 'resources' on Instance uuid 06886222-5f7f-482e-b5ee-afd7326f2c70 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.464613] env[61995]: INFO nova.scheduler.client.report [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted allocations for instance d9c53806-68c0-4872-a262-36bc05573674 [ 1067.501492] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795218, 'name': Destroy_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.574090] env[61995]: DEBUG nova.compute.manager [req-3bc483e8-1f57-49e7-b988-31dbca2bdfd8 req-ab0f16c2-aaf0-463b-87f8-66f49fdafeea service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Received event network-vif-plugged-6d70ab82-17c8-4f87-a1da-012146ad2eee {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1067.574090] env[61995]: DEBUG oslo_concurrency.lockutils [req-3bc483e8-1f57-49e7-b988-31dbca2bdfd8 req-ab0f16c2-aaf0-463b-87f8-66f49fdafeea service nova] Acquiring lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.574090] env[61995]: DEBUG oslo_concurrency.lockutils [req-3bc483e8-1f57-49e7-b988-31dbca2bdfd8 req-ab0f16c2-aaf0-463b-87f8-66f49fdafeea service nova] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.574496] env[61995]: DEBUG oslo_concurrency.lockutils [req-3bc483e8-1f57-49e7-b988-31dbca2bdfd8 req-ab0f16c2-aaf0-463b-87f8-66f49fdafeea service nova] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.574913] env[61995]: DEBUG nova.compute.manager [req-3bc483e8-1f57-49e7-b988-31dbca2bdfd8 req-ab0f16c2-aaf0-463b-87f8-66f49fdafeea service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] No waiting events found dispatching network-vif-plugged-6d70ab82-17c8-4f87-a1da-012146ad2eee {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1067.575298] env[61995]: WARNING nova.compute.manager [req-3bc483e8-1f57-49e7-b988-31dbca2bdfd8 req-ab0f16c2-aaf0-463b-87f8-66f49fdafeea service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Received unexpected event network-vif-plugged-6d70ab82-17c8-4f87-a1da-012146ad2eee for instance with vm_state building and task_state spawning. [ 1067.644353] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "d2f84704-7d96-4ed6-835f-7176ff10148a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.648017] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.653828] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795219, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.665114] env[61995]: DEBUG nova.network.neutron [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Successfully updated port: 6d70ab82-17c8-4f87-a1da-012146ad2eee {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.804045] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] In vmwareapi:vmops:_destroy_instance, exception while deleting the VM contents from the disk: oslo_vmware.exceptions.CannotDeleteFileException: Cannot delete file [datastore1] ff07d5b8-38b4-48d4-a6d0-632d37d3a91a [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Traceback (most recent call last): [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1114, in _destroy_instance [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] ds_util.file_delete(self._session, [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/nova/nova/virt/vmwareapi/ds_util.py", line 219, in file_delete [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] session._wait_for_task(file_delete_task) [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] return self.wait_for_task(task_ref) [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 398, in wait_for_task [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] return evt.wait() [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] result = hub.switch() [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] return self.greenlet.switch() [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] self.f(*self.args, **self.kw) [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 448, in _poll_task [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] raise exceptions.translate_fault(task_info.error) [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] oslo_vmware.exceptions.CannotDeleteFileException: Cannot delete file [datastore1] ff07d5b8-38b4-48d4-a6d0-632d37d3a91a [ 1067.804045] env[61995]: ERROR nova.virt.vmwareapi.vmops [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] [ 1067.804669] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1067.804669] env[61995]: INFO nova.compute.manager [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1067.804822] env[61995]: DEBUG oslo.service.loopingcall [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.805048] env[61995]: DEBUG nova.compute.manager [-] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1067.805170] env[61995]: DEBUG nova.network.neutron [-] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1067.853019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.853401] env[61995]: DEBUG nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Instance network_info: |[{"id": "2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7", "address": "fa:16:3e:dd:33:25", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ebf2f2d-13", "ovs_interfaceid": "2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1067.853841] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:33:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbd7899c-c96e-47fc-9141-5803b646917a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.861222] env[61995]: DEBUG oslo.service.loopingcall [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.861449] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1067.861679] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6323a834-fe30-45e0-8b55-518d6b6073c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.882504] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.882504] env[61995]: value = "task-795223" [ 1067.882504] env[61995]: _type = "Task" [ 1067.882504] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.890892] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795223, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.930795] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795222, 'name': ReconfigVM_Task, 'duration_secs': 0.510361} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.931118] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Reconfigured VM instance instance-00000062 to attach disk [datastore2] volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7/volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.936091] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8aa2d17-4fe5-4dd7-84f9-0b2df6f5fe4f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.949692] env[61995]: DEBUG nova.compute.utils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.955302] env[61995]: DEBUG nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1067.955489] env[61995]: DEBUG nova.network.neutron [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1067.957789] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1067.957789] env[61995]: value = "task-795224" [ 1067.957789] env[61995]: _type = "Task" [ 1067.957789] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.966796] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.973698] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc42f516-ebfd-4206-be25-30a78ae7f70c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "d9c53806-68c0-4872-a262-36bc05573674" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.946s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.001877] env[61995]: DEBUG oslo_vmware.api [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795218, 'name': Destroy_Task, 'duration_secs': 1.381096} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.004612] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Destroyed the VM [ 1068.004917] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1068.006937] env[61995]: DEBUG nova.policy [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1068.008465] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0ca67710-73ae-4a22-933d-3892b7b9285c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.013488] env[61995]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1068.014239] env[61995]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61995) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1068.014239] env[61995]: DEBUG nova.compute.utils [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Cleaning up image f52129c4-0d21-41df-aad0-db5e438b6b75 {{(pid=61995) delete_image /opt/stack/nova/nova/compute/utils.py:1322}} [ 1068.150846] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795219, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.156984] env[61995]: DEBUG nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1068.168104] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "refresh_cache-85c56f36-a858-4b81-978d-bcd4f09d2ad8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.168456] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "refresh_cache-85c56f36-a858-4b81-978d-bcd4f09d2ad8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.168596] env[61995]: DEBUG nova.network.neutron [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.258401] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee4f4f7-2227-48fb-a63f-6af2b3d83297 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.265543] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af5ea40-05a0-4962-8bb9-4a3af97adfc3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.299112] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df751c0-6cec-4f4c-93de-6a0192ab533c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.306905] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b89774-5a71-4abb-9d92-fb9c37f4c3ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.320809] env[61995]: DEBUG nova.compute.provider_tree [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.392631] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795223, 'name': CreateVM_Task, 'duration_secs': 0.394858} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.392853] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1068.393487] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.393642] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.393947] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1068.394223] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7992caaa-77da-405b-bd0d-1e2bc5953fff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.399831] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1068.399831] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52670f6e-34a8-575e-9cd3-4f3d167d40ff" [ 1068.399831] env[61995]: _type = "Task" [ 1068.399831] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.411344] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52670f6e-34a8-575e-9cd3-4f3d167d40ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.416116] env[61995]: DEBUG nova.network.neutron [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Successfully created port: 3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1068.459340] env[61995]: DEBUG nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1068.475021] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795224, 'name': ReconfigVM_Task, 'duration_secs': 0.129327} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.475021] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185404', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'name': 'volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac81c31d-d310-42c7-a6b4-166f3e82cf56', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'serial': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1068.475021] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6a61a6a-467b-4244-b4fc-e78a5e198692 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.479671] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1068.479671] env[61995]: value = "task-795225" [ 1068.479671] env[61995]: _type = "Task" [ 1068.479671] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.487430] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795225, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.572190] env[61995]: DEBUG nova.network.neutron [-] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.647837] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795219, 'name': ReconfigVM_Task, 'duration_secs': 1.142579} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.648486] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48/c1739132-aff3-4644-9522-fc02cf765bc8-rescue.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1068.649771] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a3c206-c31f-48f9-9562-5660642bed96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.688130] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f075b7b-b802-4068-a518-085417d9b456 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.715064] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1068.715064] env[61995]: value = "task-795226" [ 1068.715064] env[61995]: _type = "Task" [ 1068.715064] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.721914] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.732748] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795226, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.751031] env[61995]: DEBUG nova.network.neutron [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1068.823521] env[61995]: DEBUG nova.scheduler.client.report [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.911091] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52670f6e-34a8-575e-9cd3-4f3d167d40ff, 'name': SearchDatastore_Task, 'duration_secs': 0.01305} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.911433] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.911679] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.911922] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.912085] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.912310] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.914932] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8b72bce-1256-455b-92de-39d7094c19a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.924202] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.924623] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1068.925234] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c78488c8-9860-47b8-9c02-e8d9ec175163 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.930902] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1068.930902] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]524521de-1e28-f94f-69cd-488b6d6ab3fc" [ 1068.930902] env[61995]: _type = "Task" [ 1068.930902] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.938342] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524521de-1e28-f94f-69cd-488b6d6ab3fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.952866] env[61995]: DEBUG nova.network.neutron [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Updating instance_info_cache with network_info: [{"id": "6d70ab82-17c8-4f87-a1da-012146ad2eee", "address": "fa:16:3e:75:f3:51", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d70ab82-17", "ovs_interfaceid": "6d70ab82-17c8-4f87-a1da-012146ad2eee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.989671] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795225, 'name': Rename_Task, 'duration_secs': 0.154778} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.989964] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1068.990345] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77e79dac-ba99-445c-ae47-b1e2782d7d7f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.999034] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1068.999034] env[61995]: value = "task-795227" [ 1068.999034] env[61995]: _type = "Task" [ 1068.999034] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.005162] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795227, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.074433] env[61995]: INFO nova.compute.manager [-] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Took 1.27 seconds to deallocate network for instance. [ 1069.227060] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795226, 'name': ReconfigVM_Task, 'duration_secs': 0.403285} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.227433] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1069.227713] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2bdb194e-5b8a-4504-a273-221eabd50ad6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.233821] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1069.233821] env[61995]: value = "task-795228" [ 1069.233821] env[61995]: _type = "Task" [ 1069.233821] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.242165] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.328409] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.881s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.331386] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.610s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.332999] env[61995]: INFO nova.compute.claims [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.353184] env[61995]: INFO nova.scheduler.client.report [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Deleted allocations for instance 06886222-5f7f-482e-b5ee-afd7326f2c70 [ 1069.442059] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524521de-1e28-f94f-69cd-488b6d6ab3fc, 'name': SearchDatastore_Task, 'duration_secs': 0.015216} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.442396] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85ab2bf8-9886-4692-b61a-cf8f42ee8686 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.447879] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1069.447879] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525810b8-de74-b101-470c-b7aa791de29f" [ 1069.447879] env[61995]: _type = "Task" [ 1069.447879] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.456458] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "refresh_cache-85c56f36-a858-4b81-978d-bcd4f09d2ad8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.456809] env[61995]: DEBUG nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Instance network_info: |[{"id": "6d70ab82-17c8-4f87-a1da-012146ad2eee", "address": "fa:16:3e:75:f3:51", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d70ab82-17", "ovs_interfaceid": "6d70ab82-17c8-4f87-a1da-012146ad2eee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.457141] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525810b8-de74-b101-470c-b7aa791de29f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.457541] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:f3:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d70ab82-17c8-4f87-a1da-012146ad2eee', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.464853] env[61995]: DEBUG oslo.service.loopingcall [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.465385] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1069.465611] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1002c40e-2ecd-4f1a-80c2-39cc08058a8d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.481462] env[61995]: DEBUG nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1069.488894] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.488894] env[61995]: value = "task-795229" [ 1069.488894] env[61995]: _type = "Task" [ 1069.488894] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.496749] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795229, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.505949] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795227, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.509801] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.510062] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.510292] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.510660] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.510785] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.510998] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.511268] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.511445] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.511612] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.511778] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.511949] env[61995]: DEBUG nova.virt.hardware [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.512925] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0567bf-2767-4e32-a2fe-77d2f9e7482f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.522040] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b749ca-9a6f-4c24-ae45-f6b4242599d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.538262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.581854] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.590061] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.590176] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.604313] env[61995]: DEBUG nova.compute.manager [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Received event network-changed-6d70ab82-17c8-4f87-a1da-012146ad2eee {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1069.604550] env[61995]: DEBUG nova.compute.manager [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Refreshing instance network info cache due to event network-changed-6d70ab82-17c8-4f87-a1da-012146ad2eee. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1069.604767] env[61995]: DEBUG oslo_concurrency.lockutils [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] Acquiring lock "refresh_cache-85c56f36-a858-4b81-978d-bcd4f09d2ad8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.604914] env[61995]: DEBUG oslo_concurrency.lockutils [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] Acquired lock "refresh_cache-85c56f36-a858-4b81-978d-bcd4f09d2ad8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.605094] env[61995]: DEBUG nova.network.neutron [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Refreshing network info cache for port 6d70ab82-17c8-4f87-a1da-012146ad2eee {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1069.744890] env[61995]: DEBUG oslo_vmware.api [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795228, 'name': PowerOnVM_Task, 'duration_secs': 0.457875} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.745194] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1069.747919] env[61995]: DEBUG nova.compute.manager [None req-57f9c0d3-0481-4e1f-b21d-f8fdd6f7296b tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.748751] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd376a6d-53de-4854-82f0-8305dbbc7a05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.861513] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b82462f6-786e-4740-82a5-1f2f8f7b050a tempest-ListServerFiltersTestJSON-698967261 tempest-ListServerFiltersTestJSON-698967261-project-member] Lock "06886222-5f7f-482e-b5ee-afd7326f2c70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.313s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.957867] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525810b8-de74-b101-470c-b7aa791de29f, 'name': SearchDatastore_Task, 'duration_secs': 0.014251} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.958254] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.958565] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] a73742da-3d5c-4b71-8e92-32e26d404f37/a73742da-3d5c-4b71-8e92-32e26d404f37.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1069.958832] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7eb33157-e8ef-4e28-830c-593de8da43c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.964901] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1069.964901] env[61995]: value = "task-795230" [ 1069.964901] env[61995]: _type = "Task" [ 1069.964901] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.972579] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.998029] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795229, 'name': CreateVM_Task, 'duration_secs': 0.363044} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.000964] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1070.002056] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.002056] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.002230] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.002916] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7488c415-b733-463f-86f7-79a0d2b73257 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.007764] env[61995]: DEBUG oslo_vmware.api [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795227, 'name': PowerOnVM_Task, 'duration_secs': 0.558008} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.008948] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1070.009210] env[61995]: INFO nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 7.27 seconds to spawn the instance on the hypervisor. [ 1070.009367] env[61995]: DEBUG nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.009688] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1070.009688] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dd1bad-cc9d-e36f-0db6-b7665322a18b" [ 1070.009688] env[61995]: _type = "Task" [ 1070.009688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.010358] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a7c50c-d104-4c3d-ac43-3a977c8d92b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.030383] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52dd1bad-cc9d-e36f-0db6-b7665322a18b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.085131] env[61995]: DEBUG nova.compute.manager [req-1b65411a-40bb-4adf-a0c8-c761a87e7869 req-08f3c7e5-9f0f-4a9a-8074-2338a051b70b service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-vif-plugged-3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1070.085363] env[61995]: DEBUG oslo_concurrency.lockutils [req-1b65411a-40bb-4adf-a0c8-c761a87e7869 req-08f3c7e5-9f0f-4a9a-8074-2338a051b70b service nova] Acquiring lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.085565] env[61995]: DEBUG oslo_concurrency.lockutils [req-1b65411a-40bb-4adf-a0c8-c761a87e7869 req-08f3c7e5-9f0f-4a9a-8074-2338a051b70b service nova] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.086559] env[61995]: DEBUG oslo_concurrency.lockutils [req-1b65411a-40bb-4adf-a0c8-c761a87e7869 req-08f3c7e5-9f0f-4a9a-8074-2338a051b70b service nova] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.086559] env[61995]: DEBUG nova.compute.manager [req-1b65411a-40bb-4adf-a0c8-c761a87e7869 req-08f3c7e5-9f0f-4a9a-8074-2338a051b70b service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] No waiting events found dispatching network-vif-plugged-3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.086559] env[61995]: WARNING nova.compute.manager [req-1b65411a-40bb-4adf-a0c8-c761a87e7869 req-08f3c7e5-9f0f-4a9a-8074-2338a051b70b service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received unexpected event network-vif-plugged-3cd23361-0e98-4153-b3ea-06ce74356da6 for instance with vm_state building and task_state spawning. [ 1070.092435] env[61995]: DEBUG nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1070.195309] env[61995]: DEBUG nova.network.neutron [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Successfully updated port: 3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1070.476152] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795230, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.526290] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52dd1bad-cc9d-e36f-0db6-b7665322a18b, 'name': SearchDatastore_Task, 'duration_secs': 0.019456} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.529915] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.530220] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.530498] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.530664] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.530855] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.542484] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4f1f93b-7479-4148-ada0-04c33a0d437f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.546478] env[61995]: INFO nova.compute.manager [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 18.49 seconds to build instance. [ 1070.558457] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.558702] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1070.559654] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c519eec0-c0e4-4982-818c-c6fc17d267fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.572142] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1070.572142] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b3507f-31f3-8749-a2da-f14b363c7e10" [ 1070.572142] env[61995]: _type = "Task" [ 1070.572142] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.580019] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b3507f-31f3-8749-a2da-f14b363c7e10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.583794] env[61995]: DEBUG nova.network.neutron [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Updated VIF entry in instance network info cache for port 6d70ab82-17c8-4f87-a1da-012146ad2eee. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1070.584116] env[61995]: DEBUG nova.network.neutron [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Updating instance_info_cache with network_info: [{"id": "6d70ab82-17c8-4f87-a1da-012146ad2eee", "address": "fa:16:3e:75:f3:51", "network": {"id": "38dc67a2-9a7a-47dd-ad28-9c6613fac173", "bridge": "br-int", "label": "tempest-ServersTestJSON-478563178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2018aae65cf4bf98b5ac82519201cf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d70ab82-17", "ovs_interfaceid": "6d70ab82-17c8-4f87-a1da-012146ad2eee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.619828] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.646033] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333f7315-32a7-4644-8fcc-dbd9a12cb2f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.654591] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b291ae1a-c710-45a0-ad8e-b4110455977e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.687148] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e91675-8f02-4fdd-80bd-cb8f37324924 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.697699] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcc1a9a-182d-447b-9612-ff28ffe1acb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.702852] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.703135] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.703392] env[61995]: DEBUG nova.network.neutron [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1070.715608] env[61995]: DEBUG nova.compute.provider_tree [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.976022] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795230, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648723} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.976298] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] a73742da-3d5c-4b71-8e92-32e26d404f37/a73742da-3d5c-4b71-8e92-32e26d404f37.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1070.976515] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1070.976764] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f5289da-bd56-4823-9e95-29511c06da4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.982888] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1070.982888] env[61995]: value = "task-795231" [ 1070.982888] env[61995]: _type = "Task" [ 1070.982888] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.991131] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795231, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.049453] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6f328f86-41e1-4213-ab1c-e98a203a1bbf tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.015s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.080562] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b3507f-31f3-8749-a2da-f14b363c7e10, 'name': SearchDatastore_Task, 'duration_secs': 0.02721} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.081386] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba19ba20-33bd-4ad1-a138-e76e63d45069 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.086773] env[61995]: DEBUG oslo_concurrency.lockutils [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] Releasing lock "refresh_cache-85c56f36-a858-4b81-978d-bcd4f09d2ad8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.087035] env[61995]: DEBUG nova.compute.manager [req-e11834b7-dca5-4c93-a4ad-e52f6ad5fcee req-7655888a-3110-48ea-ba6f-d23fed769ba1 service nova] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Received event network-vif-deleted-913513ba-6ddd-42c3-a202-27978a173b14 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1071.087535] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1071.087535] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ae2d03-f3f8-d4f3-1135-3e93e5cecfb4" [ 1071.087535] env[61995]: _type = "Task" [ 1071.087535] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.095577] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ae2d03-f3f8-d4f3-1135-3e93e5cecfb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.193544] env[61995]: INFO nova.compute.manager [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Unrescuing [ 1071.193915] env[61995]: DEBUG oslo_concurrency.lockutils [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.194140] env[61995]: DEBUG oslo_concurrency.lockutils [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquired lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.194365] env[61995]: DEBUG nova.network.neutron [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1071.218477] env[61995]: DEBUG nova.scheduler.client.report [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.245314] env[61995]: DEBUG nova.network.neutron [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1071.407526] env[61995]: DEBUG nova.network.neutron [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.498395] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795231, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.194539} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.499085] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.500635] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41019ac5-21fe-4d27-9f8c-cb0296151a76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.526018] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] a73742da-3d5c-4b71-8e92-32e26d404f37/a73742da-3d5c-4b71-8e92-32e26d404f37.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.526556] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f53ba96f-4fa1-4820-876c-e51d87694c42 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.552078] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1071.552078] env[61995]: value = "task-795232" [ 1071.552078] env[61995]: _type = "Task" [ 1071.552078] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.561675] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795232, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.600469] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ae2d03-f3f8-d4f3-1135-3e93e5cecfb4, 'name': SearchDatastore_Task, 'duration_secs': 0.045912} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.600469] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.600469] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 85c56f36-a858-4b81-978d-bcd4f09d2ad8/85c56f36-a858-4b81-978d-bcd4f09d2ad8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1071.600469] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57fc2036-2eae-432d-9973-e03f85936b2c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.609156] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1071.609156] env[61995]: value = "task-795233" [ 1071.609156] env[61995]: _type = "Task" [ 1071.609156] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.615433] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.723094] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.723682] env[61995]: DEBUG nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1071.728983] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.191s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.909879] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.911908] env[61995]: DEBUG nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Instance network_info: |[{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1071.912776] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:91:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cd23361-0e98-4153-b3ea-06ce74356da6', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.921178] env[61995]: DEBUG oslo.service.loopingcall [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.921488] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1071.921731] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a27d0af-191a-4e49-bc95-267a39249fdd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.953024] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.953024] env[61995]: value = "task-795234" [ 1071.953024] env[61995]: _type = "Task" [ 1071.953024] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.966010] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795234, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.036560] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474e618d-d374-4780-9fa7-acd21a863112 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.046062] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754377fe-61c9-478b-8a56-e96818a691a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.051040] env[61995]: DEBUG nova.network.neutron [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.087788] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffc37be-a0b5-41f2-85b5-b71722d4afc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.095044] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795232, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.101845] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fbc706-651c-424c-9ccc-aa6f289aaa67 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.120556] env[61995]: DEBUG nova.compute.provider_tree [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.127524] env[61995]: DEBUG nova.compute.manager [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1072.127759] env[61995]: DEBUG nova.compute.manager [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing instance network info cache due to event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1072.128016] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.128315] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.128542] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1072.134097] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795233, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.234100] env[61995]: DEBUG nova.compute.utils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1072.236199] env[61995]: DEBUG nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1072.236474] env[61995]: DEBUG nova.network.neutron [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1072.307511] env[61995]: DEBUG nova.policy [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '671c956f9796411e9b66268420075628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26c7cc2bdcb6462d8154127098e94875', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1072.463526] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795234, 'name': CreateVM_Task, 'duration_secs': 0.422191} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.463738] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1072.464491] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.464670] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.465022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1072.465292] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe8f2bd0-3c22-4c86-a411-2606ed5e0244 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.470703] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1072.470703] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52aea0a1-cb01-c8c5-ec3e-79cab053cc2f" [ 1072.470703] env[61995]: _type = "Task" [ 1072.470703] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.478396] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52aea0a1-cb01-c8c5-ec3e-79cab053cc2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.557439] env[61995]: DEBUG oslo_concurrency.lockutils [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Releasing lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.558159] env[61995]: DEBUG nova.objects.instance [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lazy-loading 'flavor' on Instance uuid 66c56f7b-1fc7-4aed-8afc-350817e1ca48 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.571778] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795232, 'name': ReconfigVM_Task, 'duration_secs': 0.580544} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.572157] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfigured VM instance instance-00000063 to attach disk [datastore2] a73742da-3d5c-4b71-8e92-32e26d404f37/a73742da-3d5c-4b71-8e92-32e26d404f37.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.573454] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c612f85-1c54-4f97-9cb5-c8fe3b815175 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.578930] env[61995]: DEBUG nova.network.neutron [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Successfully created port: e697de59-3c5b-4928-a1df-9bf03e753d3c {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1072.582710] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1072.582710] env[61995]: value = "task-795235" [ 1072.582710] env[61995]: _type = "Task" [ 1072.582710] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.594097] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795235, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.620023] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59752} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.620023] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 85c56f36-a858-4b81-978d-bcd4f09d2ad8/85c56f36-a858-4b81-978d-bcd4f09d2ad8.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1072.620023] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.620023] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5dd7bb49-c1ff-49f9-a663-0c257ea72e35 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.627025] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1072.627025] env[61995]: value = "task-795236" [ 1072.627025] env[61995]: _type = "Task" [ 1072.627025] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.633508] env[61995]: DEBUG nova.scheduler.client.report [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.637348] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795236, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.740595] env[61995]: DEBUG nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1072.947406] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updated VIF entry in instance network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1072.947406] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.983603] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52aea0a1-cb01-c8c5-ec3e-79cab053cc2f, 'name': SearchDatastore_Task, 'duration_secs': 0.010898} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.983603] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.983603] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.983603] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.983603] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.983603] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.983603] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60c29dcf-7b2c-4c9a-8ae2-d2212bb2830b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.999164] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.999408] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1073.000193] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76a6ec9d-0638-490a-9abb-49a5662d7622 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.006991] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1073.006991] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526fb46a-146e-c4b8-6fe5-c5bd5290e7f2" [ 1073.006991] env[61995]: _type = "Task" [ 1073.006991] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.013871] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526fb46a-146e-c4b8-6fe5-c5bd5290e7f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.076563] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9804d4ea-b388-419e-aa56-4952cf32d08c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.103530] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1073.108622] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-078c31f0-7fd4-4f2a-a025-a95930a4ab2b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.115914] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795235, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.120993] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1073.120993] env[61995]: value = "task-795237" [ 1073.120993] env[61995]: _type = "Task" [ 1073.120993] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.132146] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.137151] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795236, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082985} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.137151] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.137953] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05978a03-6120-4cf8-8e8d-0dd80734652e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.141496] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.413s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.141824] env[61995]: INFO nova.compute.manager [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Successfully reverted task state from image_uploading on failure for instance. [ 1073.144614] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.563s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.144960] env[61995]: DEBUG nova.objects.instance [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lazy-loading 'resources' on Instance uuid ff07d5b8-38b4-48d4-a6d0-632d37d3a91a {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server [None req-2819985c-53af-4089-b1d7-a55c47ae2c10 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Exception during message handling: oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.vm.Snapshot:snapshot-185406' has already been deleted or has not been completely created [ 1073.148435] env[61995]: Cause: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-185406' has already been deleted or has not been completely created' [ 1073.148435] env[61995]: Faults: [ManagedObjectNotFound] [ 1073.148435] env[61995]: Details: {'obj': 'snapshot-185406'} [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server response = request(managed_object, **kwargs) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__ [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server return client.invoke(args, kwargs) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server result = self.send(soapenv, timeout=timeout) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server return self.process_reply(reply.message, None, None) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server raise WebFault(fault, replyroot) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server suds.WebFault: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-185406' has already been deleted or has not been completely created' [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server return api_method(*args, **kwargs) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server raise exceptions.VimFaultException(fault_list, fault_string, [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server oslo_vmware.exceptions.VimFaultException: The object 'vim.vm.Snapshot:snapshot-185406' has already been deleted or has not been completely created [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server Cause: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-185406' has already been deleted or has not been completely created' [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server Faults: [ManagedObjectNotFound] [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server Details: {'obj': 'snapshot-185406'} [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1073.148435] env[61995]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 233, in decorated_function [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 230, in decorated_function [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return function(self, context, image_id, instance, [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4442, in snapshot_instance [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server self._snapshot_instance(context, image_id, instance, [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4475, in _snapshot_instance [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server self.driver.snapshot(context, instance, image_id, [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 565, in snapshot [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server self._vmops.snapshot(context, instance, image_id, update_task_state) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1034, in snapshot [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server self._delete_vm_snapshot(instance, vm_ref, snapshot_ref) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/decorator.py", line 232, in fun [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return caller(func, *(extras + args), **kw) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 124, in retry_if_task_in_progress [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server f(*args, **kwargs) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 933, in _delete_vm_snapshot [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server delete_snapshot_task = self._session._call_method( [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 127, in _call_method [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception() as ctxt: [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 125, in _call_method [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return self.invoke_api(module, method, *args, **kwargs) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return _invoke_api(module, method, *args, **kwargs) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return evt.wait() [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server result = hub.switch() [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server return self.greenlet.switch() [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server idle = self.f(*self.args, **self.kw) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server result = f(*args, **kwargs) [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api [ 1073.150429] env[61995]: ERROR oslo_messaging.rpc.server raise clazz(str(excep), [ 1073.151826] env[61995]: ERROR oslo_messaging.rpc.server oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.vm.Snapshot:snapshot-185406' has already been deleted or has not been completely created [ 1073.151826] env[61995]: ERROR oslo_messaging.rpc.server Cause: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-185406' has already been deleted or has not been completely created' [ 1073.151826] env[61995]: ERROR oslo_messaging.rpc.server Faults: [ManagedObjectNotFound] [ 1073.151826] env[61995]: ERROR oslo_messaging.rpc.server Details: {'obj': 'snapshot-185406'} [ 1073.151826] env[61995]: ERROR oslo_messaging.rpc.server [ 1073.169250] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 85c56f36-a858-4b81-978d-bcd4f09d2ad8/85c56f36-a858-4b81-978d-bcd4f09d2ad8.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.173022] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-199954b9-23e4-4d66-a93a-3faf356a5b97 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.191263] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1073.191263] env[61995]: value = "task-795238" [ 1073.191263] env[61995]: _type = "Task" [ 1073.191263] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.199464] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795238, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.230184] env[61995]: DEBUG nova.compute.manager [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1073.448587] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.448925] env[61995]: DEBUG nova.compute.manager [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Received event network-changed-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1073.449155] env[61995]: DEBUG nova.compute.manager [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Refreshing instance network info cache due to event network-changed-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1073.449392] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Acquiring lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.449541] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Acquired lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.449711] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Refreshing network info cache for port ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1073.515868] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526fb46a-146e-c4b8-6fe5-c5bd5290e7f2, 'name': SearchDatastore_Task, 'duration_secs': 0.044037} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.516658] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8ded912-ba1d-4cb5-8ff5-fdfe7c001e67 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.521775] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1073.521775] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527165c6-8036-23a6-9868-ee2b58f3f410" [ 1073.521775] env[61995]: _type = "Task" [ 1073.521775] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.529437] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527165c6-8036-23a6-9868-ee2b58f3f410, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.612065] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795235, 'name': Rename_Task, 'duration_secs': 0.556805} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.612374] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1073.612593] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9db54831-340f-4fd0-8a11-2158fabb5032 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.618566] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1073.618566] env[61995]: value = "task-795239" [ 1073.618566] env[61995]: _type = "Task" [ 1073.618566] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.628980] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795239, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.632345] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.702558] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795238, 'name': ReconfigVM_Task, 'duration_secs': 0.300329} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.702856] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 85c56f36-a858-4b81-978d-bcd4f09d2ad8/85c56f36-a858-4b81-978d-bcd4f09d2ad8.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.703513] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed36fbcb-cc6e-4fa8-99ce-6acc5a843720 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.709803] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1073.709803] env[61995]: value = "task-795240" [ 1073.709803] env[61995]: _type = "Task" [ 1073.709803] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.717138] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795240, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.745114] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.752440] env[61995]: DEBUG nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1073.777654] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.777995] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.778234] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.778445] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.778599] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.778751] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.778968] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.779183] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.779379] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.779551] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.779731] env[61995]: DEBUG nova.virt.hardware [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.780580] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7061a66-4f68-4382-adee-5dad76dc777e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.788886] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c311c5fc-6276-4a3e-bdbc-aefeddd42936 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.858273] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8da746-90fb-4c5b-81fd-fdc3f87362ca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.865469] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67eed1e7-06ab-49f3-be2c-264bc16c673d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.893797] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e89abe6-7a31-446f-bf7f-77acf3562639 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.900729] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de01470-8a52-4945-8606-d8e295a59c2b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.913825] env[61995]: DEBUG nova.compute.provider_tree [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.034494] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527165c6-8036-23a6-9868-ee2b58f3f410, 'name': SearchDatastore_Task, 'duration_secs': 0.08963} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.034950] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.035262] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 50717fa4-956c-4e59-ab07-e6e8ec6f16bf/50717fa4-956c-4e59-ab07-e6e8ec6f16bf.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1074.035545] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a33e4458-e66a-4be6-b132-e9f917d2d0da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.041833] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1074.041833] env[61995]: value = "task-795241" [ 1074.041833] env[61995]: _type = "Task" [ 1074.041833] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.050327] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795241, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.132298] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795239, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.135848] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.199622] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updated VIF entry in instance network info cache for port ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1074.200234] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updating instance_info_cache with network_info: [{"id": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "address": "fa:16:3e:d6:67:26", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcee8fc-e0", "ovs_interfaceid": "ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.221984] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795240, 'name': Rename_Task, 'duration_secs': 0.150883} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.222445] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1074.222775] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2221f12-0259-4d30-91d3-f0d79871425e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.230334] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1074.230334] env[61995]: value = "task-795242" [ 1074.230334] env[61995]: _type = "Task" [ 1074.230334] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.247041] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.417344] env[61995]: DEBUG nova.scheduler.client.report [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.434750] env[61995]: DEBUG nova.compute.manager [req-5eef1b04-b903-4bb6-9b0d-3969ca9e9d6d req-cd8f69bd-11e0-4d73-8388-7421ef9e4957 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Received event network-vif-plugged-e697de59-3c5b-4928-a1df-9bf03e753d3c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1074.435036] env[61995]: DEBUG oslo_concurrency.lockutils [req-5eef1b04-b903-4bb6-9b0d-3969ca9e9d6d req-cd8f69bd-11e0-4d73-8388-7421ef9e4957 service nova] Acquiring lock "d2f84704-7d96-4ed6-835f-7176ff10148a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.435339] env[61995]: DEBUG oslo_concurrency.lockutils [req-5eef1b04-b903-4bb6-9b0d-3969ca9e9d6d req-cd8f69bd-11e0-4d73-8388-7421ef9e4957 service nova] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.435521] env[61995]: DEBUG oslo_concurrency.lockutils [req-5eef1b04-b903-4bb6-9b0d-3969ca9e9d6d req-cd8f69bd-11e0-4d73-8388-7421ef9e4957 service nova] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.436313] env[61995]: DEBUG nova.compute.manager [req-5eef1b04-b903-4bb6-9b0d-3969ca9e9d6d req-cd8f69bd-11e0-4d73-8388-7421ef9e4957 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] No waiting events found dispatching network-vif-plugged-e697de59-3c5b-4928-a1df-9bf03e753d3c {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1074.436313] env[61995]: WARNING nova.compute.manager [req-5eef1b04-b903-4bb6-9b0d-3969ca9e9d6d req-cd8f69bd-11e0-4d73-8388-7421ef9e4957 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Received unexpected event network-vif-plugged-e697de59-3c5b-4928-a1df-9bf03e753d3c for instance with vm_state building and task_state spawning. [ 1074.552205] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795241, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.568899] env[61995]: DEBUG nova.network.neutron [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Successfully updated port: e697de59-3c5b-4928-a1df-9bf03e753d3c {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1074.633140] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795237, 'name': PowerOffVM_Task, 'duration_secs': 1.17474} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.636630] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1074.641991] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfiguring VM instance instance-00000058 to detach disk 2002 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1074.642290] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795239, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.642524] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0bbe2e1-1450-4d31-a268-5e03f80375e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.661409] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1074.661409] env[61995]: value = "task-795243" [ 1074.661409] env[61995]: _type = "Task" [ 1074.661409] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.670814] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795243, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.703674] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Releasing lock "refresh_cache-b40b1866-ced3-40ef-9ab7-30d58b75f288" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.703986] env[61995]: DEBUG nova.compute.manager [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Received event network-changed-7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1074.704216] env[61995]: DEBUG nova.compute.manager [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Refreshing instance network info cache due to event network-changed-7c06dfad-348a-49b5-8d8d-d606f4b559c9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1074.704603] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Acquiring lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.704798] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Acquired lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.705031] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Refreshing network info cache for port 7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1074.742829] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795242, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.922651] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.925785] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.305s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.926850] env[61995]: INFO nova.compute.claims [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.947822] env[61995]: INFO nova.scheduler.client.report [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted allocations for instance ff07d5b8-38b4-48d4-a6d0-632d37d3a91a [ 1075.052497] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795241, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524593} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.052802] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 50717fa4-956c-4e59-ab07-e6e8ec6f16bf/50717fa4-956c-4e59-ab07-e6e8ec6f16bf.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1075.053079] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.053409] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d675faad-e64c-48cf-ba58-c11403bcf9ec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.059713] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1075.059713] env[61995]: value = "task-795244" [ 1075.059713] env[61995]: _type = "Task" [ 1075.059713] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.067384] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795244, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.071078] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "refresh_cache-d2f84704-7d96-4ed6-835f-7176ff10148a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.071253] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "refresh_cache-d2f84704-7d96-4ed6-835f-7176ff10148a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.071413] env[61995]: DEBUG nova.network.neutron [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1075.130016] env[61995]: DEBUG oslo_vmware.api [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795239, 'name': PowerOnVM_Task, 'duration_secs': 1.085707} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.130313] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1075.130523] env[61995]: INFO nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Took 10.41 seconds to spawn the instance on the hypervisor. [ 1075.130702] env[61995]: DEBUG nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.131495] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5670ec0-95bf-47e7-a0de-6166bc6b03cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.173071] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.241194] env[61995]: DEBUG oslo_vmware.api [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795242, 'name': PowerOnVM_Task, 'duration_secs': 0.554484} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.241561] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1075.241677] env[61995]: INFO nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Took 8.11 seconds to spawn the instance on the hypervisor. [ 1075.241857] env[61995]: DEBUG nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.242662] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47c0826-169d-42e3-a398-cbd9ad1eccdf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.459845] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4226b60b-886e-4a09-b583-1b4e3de937d2 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "ff07d5b8-38b4-48d4-a6d0-632d37d3a91a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.272s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.558715] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updated VIF entry in instance network info cache for port 7c06dfad-348a-49b5-8d8d-d606f4b559c9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1075.559347] env[61995]: DEBUG nova.network.neutron [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance_info_cache with network_info: [{"id": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "address": "fa:16:3e:33:74:c0", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c06dfad-34", "ovs_interfaceid": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.568911] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795244, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177978} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.569211] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1075.570257] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec503eca-5687-435d-b491-c9620b6b87b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.593624] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 50717fa4-956c-4e59-ab07-e6e8ec6f16bf/50717fa4-956c-4e59-ab07-e6e8ec6f16bf.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1075.594610] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68cfab87-9ecf-4ccc-8077-dfa5b95311d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.615173] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1075.615173] env[61995]: value = "task-795245" [ 1075.615173] env[61995]: _type = "Task" [ 1075.615173] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.627285] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795245, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.645671] env[61995]: DEBUG nova.network.neutron [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1075.649910] env[61995]: INFO nova.compute.manager [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Took 21.87 seconds to build instance. [ 1075.673158] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795243, 'name': ReconfigVM_Task, 'duration_secs': 0.814839} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.675488] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfigured VM instance instance-00000058 to detach disk 2002 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1075.675600] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1075.675825] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f0f434e-5e0b-459f-86fc-ab381a815fe2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.682516] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1075.682516] env[61995]: value = "task-795246" [ 1075.682516] env[61995]: _type = "Task" [ 1075.682516] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.690102] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795246, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.763123] env[61995]: INFO nova.compute.manager [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Took 19.13 seconds to build instance. [ 1075.856391] env[61995]: DEBUG nova.network.neutron [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Updating instance_info_cache with network_info: [{"id": "e697de59-3c5b-4928-a1df-9bf03e753d3c", "address": "fa:16:3e:64:76:3d", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape697de59-3c", "ovs_interfaceid": "e697de59-3c5b-4928-a1df-9bf03e753d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.065893] env[61995]: DEBUG oslo_concurrency.lockutils [req-d204977d-d5e9-448a-93b3-0b2ffabcf9f8 req-9a49ab81-ed02-4c70-bfda-ba7d5cdedffc service nova] Releasing lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.127017] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795245, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.139259] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6216ade9-aef6-4db1-b10c-10e808d27a54 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.147840] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056025ef-88d4-422c-9efe-0e48cfc58dee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.151757] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4d1b9e7b-331b-4324-a576-9be364c9731c tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.378s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.181533] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44c9e53-211b-4549-9045-ac78a76bb469 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.195819] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b671e58-8e87-4a74-9fb8-996975148806 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.199806] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795246, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.210058] env[61995]: DEBUG nova.compute.provider_tree [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.266036] env[61995]: DEBUG oslo_concurrency.lockutils [None req-af6a2d3f-ad9b-47c3-b08e-8e26a0040e31 tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.648s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.359011] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "refresh_cache-d2f84704-7d96-4ed6-835f-7176ff10148a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.359400] env[61995]: DEBUG nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Instance network_info: |[{"id": "e697de59-3c5b-4928-a1df-9bf03e753d3c", "address": "fa:16:3e:64:76:3d", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape697de59-3c", "ovs_interfaceid": "e697de59-3c5b-4928-a1df-9bf03e753d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1076.359847] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:76:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e697de59-3c5b-4928-a1df-9bf03e753d3c', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.368222] env[61995]: DEBUG oslo.service.loopingcall [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.368463] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1076.368722] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f02ff94-a2e1-4035-871d-3360dbb4b9d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.389821] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.389821] env[61995]: value = "task-795247" [ 1076.389821] env[61995]: _type = "Task" [ 1076.389821] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.397865] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795247, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.475464] env[61995]: DEBUG nova.compute.manager [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Received event network-changed-e697de59-3c5b-4928-a1df-9bf03e753d3c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1076.475910] env[61995]: DEBUG nova.compute.manager [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Refreshing instance network info cache due to event network-changed-e697de59-3c5b-4928-a1df-9bf03e753d3c. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1076.476359] env[61995]: DEBUG oslo_concurrency.lockutils [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] Acquiring lock "refresh_cache-d2f84704-7d96-4ed6-835f-7176ff10148a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.476696] env[61995]: DEBUG oslo_concurrency.lockutils [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] Acquired lock "refresh_cache-d2f84704-7d96-4ed6-835f-7176ff10148a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.478780] env[61995]: DEBUG nova.network.neutron [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Refreshing network info cache for port e697de59-3c5b-4928-a1df-9bf03e753d3c {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1076.627466] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795245, 'name': ReconfigVM_Task, 'duration_secs': 0.589714} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.627756] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 50717fa4-956c-4e59-ab07-e6e8ec6f16bf/50717fa4-956c-4e59-ab07-e6e8ec6f16bf.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1076.628428] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21e57f3c-9138-4627-ab6f-0f12c2a9ce4b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.634487] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1076.634487] env[61995]: value = "task-795248" [ 1076.634487] env[61995]: _type = "Task" [ 1076.634487] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.642354] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795248, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.693385] env[61995]: DEBUG oslo_vmware.api [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795246, 'name': PowerOnVM_Task, 'duration_secs': 0.780501} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.693723] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1076.694014] env[61995]: DEBUG nova.compute.manager [None req-991e0f5d-443a-4881-bd3d-f86300ff7b14 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.694872] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2714a245-b3a1-4609-991b-5d1ed2825953 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.714171] env[61995]: DEBUG nova.scheduler.client.report [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.899776] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795247, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.143964] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795248, 'name': Rename_Task, 'duration_secs': 0.221865} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.146503] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1077.146782] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3c147ef-7c6a-42e0-a6ad-bd544eba66af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.152956] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1077.152956] env[61995]: value = "task-795249" [ 1077.152956] env[61995]: _type = "Task" [ 1077.152956] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.163193] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795249, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.186771] env[61995]: DEBUG nova.network.neutron [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Updated VIF entry in instance network info cache for port e697de59-3c5b-4928-a1df-9bf03e753d3c. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1077.187226] env[61995]: DEBUG nova.network.neutron [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Updating instance_info_cache with network_info: [{"id": "e697de59-3c5b-4928-a1df-9bf03e753d3c", "address": "fa:16:3e:64:76:3d", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape697de59-3c", "ovs_interfaceid": "e697de59-3c5b-4928-a1df-9bf03e753d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.218648] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.219394] env[61995]: DEBUG nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1077.222684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.478s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.400784] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795247, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.442843] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.443095] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.443350] env[61995]: DEBUG nova.compute.manager [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1077.444391] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6c04ef-ba7b-4bf9-a659-a5ffc012e87e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.451941] env[61995]: DEBUG nova.compute.manager [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1077.452589] env[61995]: DEBUG nova.objects.instance [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'flavor' on Instance uuid 85c56f36-a858-4b81-978d-bcd4f09d2ad8 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.664287] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795249, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.690119] env[61995]: DEBUG oslo_concurrency.lockutils [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] Releasing lock "refresh_cache-d2f84704-7d96-4ed6-835f-7176ff10148a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.690464] env[61995]: DEBUG nova.compute.manager [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Received event network-changed-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.690666] env[61995]: DEBUG nova.compute.manager [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Refreshing instance network info cache due to event network-changed-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1077.690892] env[61995]: DEBUG oslo_concurrency.lockutils [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] Acquiring lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.691053] env[61995]: DEBUG oslo_concurrency.lockutils [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] Acquired lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.691227] env[61995]: DEBUG nova.network.neutron [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Refreshing network info cache for port 2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1077.726334] env[61995]: DEBUG nova.compute.utils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1077.730015] env[61995]: INFO nova.compute.claims [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.733660] env[61995]: DEBUG nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1077.733825] env[61995]: DEBUG nova.network.neutron [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1077.814847] env[61995]: DEBUG nova.policy [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e08719d59b456786fce984cc1508e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c3e03b3825c43ca987d60b339fc0a52', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1077.901748] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795247, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.957499] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1077.957785] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5ee2262-86ff-4d47-a129-cca94cd3c210 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.964746] env[61995]: DEBUG oslo_vmware.api [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1077.964746] env[61995]: value = "task-795250" [ 1077.964746] env[61995]: _type = "Task" [ 1077.964746] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.974548] env[61995]: DEBUG oslo_vmware.api [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.173205] env[61995]: DEBUG oslo_vmware.api [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795249, 'name': PowerOnVM_Task, 'duration_secs': 0.675388} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.173983] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1078.173983] env[61995]: INFO nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Took 8.69 seconds to spawn the instance on the hypervisor. [ 1078.173983] env[61995]: DEBUG nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.174583] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f037f2-0a51-4b85-9019-64834d0539c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.196263] env[61995]: DEBUG nova.network.neutron [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Successfully created port: 06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1078.235160] env[61995]: DEBUG nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1078.240973] env[61995]: INFO nova.compute.resource_tracker [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating resource usage from migration 8b9e9144-718f-46d0-9851-91e5d87fd2dc [ 1078.401442] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795247, 'name': CreateVM_Task, 'duration_secs': 1.599782} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.403899] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1078.408535] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.408535] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.408535] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1078.408535] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f03600dd-6c23-4fae-ae8e-53871418362c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.410489] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1078.410489] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525f0d0b-e352-851f-df11-e1bf0d6f0310" [ 1078.410489] env[61995]: _type = "Task" [ 1078.410489] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.422592] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525f0d0b-e352-851f-df11-e1bf0d6f0310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.479320] env[61995]: DEBUG oslo_vmware.api [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795250, 'name': PowerOffVM_Task, 'duration_secs': 0.381504} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.479589] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1078.479767] env[61995]: DEBUG nova.compute.manager [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.480540] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572f5ba4-d87e-4331-a73b-516c13545504 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.492261] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2c747f-9fae-4fbd-a798-fb107546cdd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.501696] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd00041-52c0-4154-ac22-203772104794 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.536031] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261bd308-da2b-42e1-a2ef-27b47152e2fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.541555] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1daf283-34f9-4571-abde-8db29a1e6db5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.556123] env[61995]: DEBUG nova.compute.provider_tree [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.603664] env[61995]: DEBUG nova.compute.manager [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Received event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1078.603879] env[61995]: DEBUG nova.compute.manager [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing instance network info cache due to event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1078.604167] env[61995]: DEBUG oslo_concurrency.lockutils [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] Acquiring lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.604325] env[61995]: DEBUG oslo_concurrency.lockutils [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] Acquired lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.604508] env[61995]: DEBUG nova.network.neutron [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1078.618264] env[61995]: DEBUG nova.network.neutron [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updated VIF entry in instance network info cache for port 2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1078.618629] env[61995]: DEBUG nova.network.neutron [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updating instance_info_cache with network_info: [{"id": "2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7", "address": "fa:16:3e:dd:33:25", "network": {"id": "802c6f9f-9b23-432f-af24-97d79bb33363", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1290802752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4de205843d14c7a91156c9b49ab7657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ebf2f2d-13", "ovs_interfaceid": "2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.695510] env[61995]: INFO nova.compute.manager [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Took 21.54 seconds to build instance. [ 1078.921612] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525f0d0b-e352-851f-df11-e1bf0d6f0310, 'name': SearchDatastore_Task, 'duration_secs': 0.011324} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.921909] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.922166] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.922409] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.922557] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.922743] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.923008] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edcab39c-5b02-41cd-bdc1-f8fd948a5c44 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.931731] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.931911] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1078.932602] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f31ef1-f7c5-4977-bd47-c44f55cbdf0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.937592] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1078.937592] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5239040f-9157-af51-32b6-d7f22e328f58" [ 1078.937592] env[61995]: _type = "Task" [ 1078.937592] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.944831] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5239040f-9157-af51-32b6-d7f22e328f58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.999022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ea4e2907-ad3a-4381-980d-afb3f501b46a tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.058756] env[61995]: DEBUG nova.scheduler.client.report [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.121010] env[61995]: DEBUG oslo_concurrency.lockutils [req-86a512ef-5ef1-476b-9051-ae78d0faacb5 req-8be4612d-9c32-412e-99f0-b283602ce3e1 service nova] Releasing lock "refresh_cache-a73742da-3d5c-4b71-8e92-32e26d404f37" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.197327] env[61995]: DEBUG oslo_concurrency.lockutils [None req-2a2106ad-857b-4936-83f1-d8f7bfba63ec tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.062s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.249782] env[61995]: DEBUG nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1079.282039] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1079.282321] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1079.282484] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1079.282669] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1079.282894] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1079.282971] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1079.283397] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1079.283579] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1079.283761] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1079.283928] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1079.284122] env[61995]: DEBUG nova.virt.hardware [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1079.284999] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488bf58d-cef1-4537-97a9-2a41b03c1023 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.297058] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59929ea-04fe-4332-bbc0-ca1a856f171e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.404779] env[61995]: DEBUG nova.network.neutron [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updated VIF entry in instance network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1079.405224] env[61995]: DEBUG nova.network.neutron [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.448420] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5239040f-9157-af51-32b6-d7f22e328f58, 'name': SearchDatastore_Task, 'duration_secs': 0.039318} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.448676] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5763dea9-c9d4-46d8-9a42-86286b77cf63 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.454105] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1079.454105] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a63c81-1f64-dbcf-ad61-be587271355d" [ 1079.454105] env[61995]: _type = "Task" [ 1079.454105] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.461447] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a63c81-1f64-dbcf-ad61-be587271355d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.565571] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.343s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.565860] env[61995]: INFO nova.compute.manager [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Migrating [ 1079.820027] env[61995]: DEBUG nova.compute.manager [req-b93ff8ea-25a9-4b48-ac51-f53b94d4b729 req-33b0665e-7c22-4c7f-a632-1d2b0d7a4c58 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Received event network-vif-plugged-06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1079.820281] env[61995]: DEBUG oslo_concurrency.lockutils [req-b93ff8ea-25a9-4b48-ac51-f53b94d4b729 req-33b0665e-7c22-4c7f-a632-1d2b0d7a4c58 service nova] Acquiring lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.820502] env[61995]: DEBUG oslo_concurrency.lockutils [req-b93ff8ea-25a9-4b48-ac51-f53b94d4b729 req-33b0665e-7c22-4c7f-a632-1d2b0d7a4c58 service nova] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.820679] env[61995]: DEBUG oslo_concurrency.lockutils [req-b93ff8ea-25a9-4b48-ac51-f53b94d4b729 req-33b0665e-7c22-4c7f-a632-1d2b0d7a4c58 service nova] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.821049] env[61995]: DEBUG nova.compute.manager [req-b93ff8ea-25a9-4b48-ac51-f53b94d4b729 req-33b0665e-7c22-4c7f-a632-1d2b0d7a4c58 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] No waiting events found dispatching network-vif-plugged-06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1079.821049] env[61995]: WARNING nova.compute.manager [req-b93ff8ea-25a9-4b48-ac51-f53b94d4b729 req-33b0665e-7c22-4c7f-a632-1d2b0d7a4c58 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Received unexpected event network-vif-plugged-06334396-e435-480b-ab0c-0da719e15bf1 for instance with vm_state building and task_state spawning. [ 1079.908186] env[61995]: DEBUG oslo_concurrency.lockutils [req-287fbc32-2ade-4298-8e18-c5ab554b2df3 req-5c3392d3-660c-4886-b145-eefe31033fc2 service nova] Releasing lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.967195] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a63c81-1f64-dbcf-ad61-be587271355d, 'name': SearchDatastore_Task, 'duration_secs': 0.010798} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.967504] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.967715] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] d2f84704-7d96-4ed6-835f-7176ff10148a/d2f84704-7d96-4ed6-835f-7176ff10148a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1079.967971] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d503f59d-39f9-435d-9002-95f4cf8fc3cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.974970] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1079.974970] env[61995]: value = "task-795251" [ 1079.974970] env[61995]: _type = "Task" [ 1079.974970] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.983639] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.066893] env[61995]: DEBUG nova.network.neutron [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Successfully updated port: 06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1080.087433] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.087433] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquired lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.087433] env[61995]: DEBUG nova.network.neutron [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1080.488654] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475507} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.488925] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] d2f84704-7d96-4ed6-835f-7176ff10148a/d2f84704-7d96-4ed6-835f-7176ff10148a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1080.490324] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1080.491709] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e981d267-10b5-4e59-82ac-d1125a9e1903 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.497601] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1080.497601] env[61995]: value = "task-795252" [ 1080.497601] env[61995]: _type = "Task" [ 1080.497601] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.506674] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795252, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.569740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.569968] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.570140] env[61995]: DEBUG nova.network.neutron [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1080.578905] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.578905] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.578905] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.579638] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.579638] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.582124] env[61995]: INFO nova.compute.manager [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Terminating instance [ 1080.583994] env[61995]: DEBUG nova.compute.manager [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1080.584214] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1080.585040] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3e9e27-66d6-4274-82af-817bb1c59ac1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.593221] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1080.593221] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-211d0417-a22c-47ca-8c9c-4b8690687ac3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.637191] env[61995]: DEBUG nova.compute.manager [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Received event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.637191] env[61995]: DEBUG nova.compute.manager [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing instance network info cache due to event network-changed-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1080.637191] env[61995]: DEBUG oslo_concurrency.lockutils [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] Acquiring lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.637856] env[61995]: DEBUG oslo_concurrency.lockutils [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] Acquired lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.637856] env[61995]: DEBUG nova.network.neutron [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Refreshing network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1080.685906] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1080.686106] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1080.686252] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleting the datastore file [datastore2] 85c56f36-a858-4b81-978d-bcd4f09d2ad8 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.686507] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29ff3010-11cd-463f-898f-bcd12ebfcbee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.692751] env[61995]: DEBUG oslo_vmware.api [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1080.692751] env[61995]: value = "task-795254" [ 1080.692751] env[61995]: _type = "Task" [ 1080.692751] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.700857] env[61995]: DEBUG oslo_vmware.api [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795254, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.924093] env[61995]: DEBUG nova.network.neutron [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance_info_cache with network_info: [{"id": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "address": "fa:16:3e:33:74:c0", "network": {"id": "2bb0d64c-c869-4c12-b5ad-46f49aae06aa", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-906176650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83064cbae17c429d8d084837635486da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c06dfad-34", "ovs_interfaceid": "7c06dfad-348a-49b5-8d8d-d606f4b559c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.007879] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795252, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075427} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.008464] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.009305] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bda054-3a25-4fcb-87e1-7d777693bd00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.033256] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] d2f84704-7d96-4ed6-835f-7176ff10148a/d2f84704-7d96-4ed6-835f-7176ff10148a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.033256] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64cc0276-423f-4c5e-a578-2af667e1e3e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.054048] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1081.054048] env[61995]: value = "task-795255" [ 1081.054048] env[61995]: _type = "Task" [ 1081.054048] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.062547] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795255, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.119200] env[61995]: DEBUG nova.network.neutron [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1081.202559] env[61995]: DEBUG oslo_vmware.api [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795254, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164185} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.203524] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.203524] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1081.206517] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1081.206517] env[61995]: INFO nova.compute.manager [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1081.206517] env[61995]: DEBUG oslo.service.loopingcall [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.206517] env[61995]: DEBUG nova.compute.manager [-] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1081.206517] env[61995]: DEBUG nova.network.neutron [-] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1081.418889] env[61995]: DEBUG nova.network.neutron [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.428151] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Releasing lock "refresh_cache-ac81c31d-d310-42c7-a6b4-166f3e82cf56" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.459298] env[61995]: DEBUG nova.network.neutron [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updated VIF entry in instance network info cache for port ff48bd88-ebef-44df-92d3-1a4b9c6ff71d. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1081.459298] env[61995]: DEBUG nova.network.neutron [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [{"id": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "address": "fa:16:3e:25:13:6c", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff48bd88-eb", "ovs_interfaceid": "ff48bd88-ebef-44df-92d3-1a4b9c6ff71d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.567033] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795255, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.858652] env[61995]: DEBUG nova.compute.manager [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Received event network-changed-06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1081.858888] env[61995]: DEBUG nova.compute.manager [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Refreshing instance network info cache due to event network-changed-06334396-e435-480b-ab0c-0da719e15bf1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1081.859372] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Acquiring lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.923135] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.923724] env[61995]: DEBUG nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Instance network_info: |[{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1081.923890] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Acquired lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.924045] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Refreshing network info cache for port 06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1081.926653] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:41:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ada35c98-01a9-4352-98e4-1d20ba31f928', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06334396-e435-480b-ab0c-0da719e15bf1', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.934767] env[61995]: DEBUG oslo.service.loopingcall [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.939110] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1081.939826] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f167053-6b23-41a1-9aef-7cfbcd726eae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.960547] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1081.960547] env[61995]: value = "task-795256" [ 1081.960547] env[61995]: _type = "Task" [ 1081.960547] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.961183] env[61995]: DEBUG oslo_concurrency.lockutils [req-09fb2b92-1386-45f0-84f1-69235d30efed req-e61b558b-0661-40bc-9ce3-c2d99d5de8ee service nova] Releasing lock "refresh_cache-66c56f7b-1fc7-4aed-8afc-350817e1ca48" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.969794] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795256, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.048381] env[61995]: DEBUG nova.network.neutron [-] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.067975] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795255, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.471765] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795256, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.550767] env[61995]: INFO nova.compute.manager [-] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Took 1.35 seconds to deallocate network for instance. [ 1082.566574] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795255, 'name': ReconfigVM_Task, 'duration_secs': 1.051332} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.566849] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Reconfigured VM instance instance-00000066 to attach disk [datastore1] d2f84704-7d96-4ed6-835f-7176ff10148a/d2f84704-7d96-4ed6-835f-7176ff10148a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.567474] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97f1eb69-c5fc-4369-8910-ab1d182281d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.573432] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1082.573432] env[61995]: value = "task-795257" [ 1082.573432] env[61995]: _type = "Task" [ 1082.573432] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.580679] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795257, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.632956] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updated VIF entry in instance network info cache for port 06334396-e435-480b-ab0c-0da719e15bf1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1082.633346] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.664966] env[61995]: DEBUG nova.compute.manager [req-bd00ec26-83ce-42b0-9933-c7a34e50865f req-a74ca98f-5199-4743-9f62-9416f4ff0724 service nova] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Received event network-vif-deleted-6d70ab82-17c8-4f87-a1da-012146ad2eee {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.948086] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b341898f-1f57-4137-87a5-e8f969a99a36 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.969017] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance 'ac81c31d-d310-42c7-a6b4-166f3e82cf56' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1082.980184] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795256, 'name': CreateVM_Task, 'duration_secs': 0.543022} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.980371] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1082.981054] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.981238] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.981544] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1082.981801] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5179ad-4d2d-4517-a651-cf5a4b3597ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.986017] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1082.986017] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527d40c8-ff96-49b1-6dfc-23106e5c5367" [ 1082.986017] env[61995]: _type = "Task" [ 1082.986017] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.993889] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527d40c8-ff96-49b1-6dfc-23106e5c5367, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.057342] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.057631] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.057858] env[61995]: DEBUG nova.objects.instance [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'resources' on Instance uuid 85c56f36-a858-4b81-978d-bcd4f09d2ad8 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.083661] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795257, 'name': Rename_Task, 'duration_secs': 0.149804} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.084394] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1083.084394] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86a72bad-9c0f-46ac-9e82-c199f6f73604 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.090127] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1083.090127] env[61995]: value = "task-795258" [ 1083.090127] env[61995]: _type = "Task" [ 1083.090127] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.097527] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795258, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.136046] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Releasing lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.136354] env[61995]: DEBUG nova.compute.manager [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1083.136529] env[61995]: DEBUG nova.compute.manager [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing instance network info cache due to event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1083.136743] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.136887] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.137062] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.478059] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1083.478518] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63ac6574-44ea-4a83-9cc8-e0ee5ee70415 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.486399] env[61995]: DEBUG oslo_vmware.api [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1083.486399] env[61995]: value = "task-795259" [ 1083.486399] env[61995]: _type = "Task" [ 1083.486399] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.500665] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527d40c8-ff96-49b1-6dfc-23106e5c5367, 'name': SearchDatastore_Task, 'duration_secs': 0.010777} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.500850] env[61995]: DEBUG oslo_vmware.api [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795259, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.501164] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.501402] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.501643] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.501789] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.502386] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.502386] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ab211c6-fca3-4249-9deb-350425e8acc3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.511217] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.511444] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1083.512246] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7710a61-59e2-4493-a86e-cc2b2e8143f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.518137] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1083.518137] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ea75ea-5c64-1645-d116-d62d853b6d97" [ 1083.518137] env[61995]: _type = "Task" [ 1083.518137] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.525621] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ea75ea-5c64-1645-d116-d62d853b6d97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.603310] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795258, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.795464] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8ba5ed-32bf-4ed4-9480-135e6454037a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.806018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce76b66-68bb-4e61-a66f-09aebc216a59 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.837116] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3159b287-cf54-4111-979b-f37b93028fa7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.845192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddc1ff6-12be-4aa2-b717-9471a26a6855 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.862593] env[61995]: DEBUG nova.compute.provider_tree [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.890125] env[61995]: DEBUG nova.compute.manager [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1083.890125] env[61995]: DEBUG nova.compute.manager [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing instance network info cache due to event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1083.890125] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.890125] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.890770] env[61995]: DEBUG nova.network.neutron [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.913113] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updated VIF entry in instance network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1083.913659] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.996740] env[61995]: DEBUG oslo_vmware.api [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795259, 'name': PowerOffVM_Task, 'duration_secs': 0.226392} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.996740] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1083.996954] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance 'ac81c31d-d310-42c7-a6b4-166f3e82cf56' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1084.029126] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ea75ea-5c64-1645-d116-d62d853b6d97, 'name': SearchDatastore_Task, 'duration_secs': 0.012467} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.029897] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f86d90d-cf9e-4e28-a8af-14807193639c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.035688] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1084.035688] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c91d4a-84ae-871a-1a5a-645f9f6b99ab" [ 1084.035688] env[61995]: _type = "Task" [ 1084.035688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.043413] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c91d4a-84ae-871a-1a5a-645f9f6b99ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.100500] env[61995]: DEBUG oslo_vmware.api [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795258, 'name': PowerOnVM_Task, 'duration_secs': 0.549715} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.100854] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1084.101057] env[61995]: INFO nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Took 10.35 seconds to spawn the instance on the hypervisor. [ 1084.101198] env[61995]: DEBUG nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1084.102060] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e8b05d-34e0-4ac3-937f-7bcd06078ab0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.220895] env[61995]: DEBUG oslo_concurrency.lockutils [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "83369251-b00f-4595-bf98-28a3ec84f037" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.221165] env[61995]: DEBUG oslo_concurrency.lockutils [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.368286] env[61995]: DEBUG nova.scheduler.client.report [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.416750] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.417116] env[61995]: DEBUG nova.compute.manager [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1084.417375] env[61995]: DEBUG nova.compute.manager [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing instance network info cache due to event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1084.417650] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.504208] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.504499] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.504672] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.504859] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.505021] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.505179] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.505390] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.506113] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.506113] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.506113] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.506113] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.511046] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b41be441-29cc-44b7-b707-af9a836f849b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.531245] env[61995]: DEBUG oslo_vmware.api [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1084.531245] env[61995]: value = "task-795260" [ 1084.531245] env[61995]: _type = "Task" [ 1084.531245] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.543275] env[61995]: DEBUG oslo_vmware.api [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795260, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.550033] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c91d4a-84ae-871a-1a5a-645f9f6b99ab, 'name': SearchDatastore_Task, 'duration_secs': 0.042978} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.550183] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.550557] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1084.550898] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ddd8100-8ce4-4a03-ac1f-d90788bce404 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.558319] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1084.558319] env[61995]: value = "task-795261" [ 1084.558319] env[61995]: _type = "Task" [ 1084.558319] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.566186] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795261, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.609460] env[61995]: DEBUG nova.network.neutron [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updated VIF entry in instance network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1084.609460] env[61995]: DEBUG nova.network.neutron [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.619960] env[61995]: INFO nova.compute.manager [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Took 15.94 seconds to build instance. [ 1084.724230] env[61995]: DEBUG nova.compute.utils [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1084.873902] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.895179] env[61995]: INFO nova.scheduler.client.report [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted allocations for instance 85c56f36-a858-4b81-978d-bcd4f09d2ad8 [ 1085.043328] env[61995]: DEBUG oslo_vmware.api [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795260, 'name': ReconfigVM_Task, 'duration_secs': 0.188617} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.043706] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance 'ac81c31d-d310-42c7-a6b4-166f3e82cf56' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1085.069083] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795261, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.112513] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.112938] env[61995]: DEBUG nova.compute.manager [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.113037] env[61995]: DEBUG nova.compute.manager [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing instance network info cache due to event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1085.113226] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.113381] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.113552] env[61995]: DEBUG nova.network.neutron [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1085.114954] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.115149] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1085.122738] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f104cebd-47ef-47f2-beb8-b8a34db327e5 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.478s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.228617] env[61995]: DEBUG oslo_concurrency.lockutils [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.404144] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0227b113-57fc-46ff-9764-9c102bdbf6df tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "85c56f36-a858-4b81-978d-bcd4f09d2ad8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.825s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.551943] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.552266] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.552462] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.552674] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.552848] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.553032] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.553264] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.553455] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.553660] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.553858] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.554100] env[61995]: DEBUG nova.virt.hardware [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.572705] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795261, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644495} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.573657] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1085.573887] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.574174] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3d35a46-f88c-4d56-96d6-5b1495241692 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.583616] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1085.583616] env[61995]: value = "task-795262" [ 1085.583616] env[61995]: _type = "Task" [ 1085.583616] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.591937] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795262, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.958949] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updated VIF entry in instance network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1085.959424] env[61995]: DEBUG nova.network.neutron [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.987802] env[61995]: DEBUG nova.network.neutron [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updated VIF entry in instance network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1085.988213] env[61995]: DEBUG nova.network.neutron [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.067169] env[61995]: ERROR nova.compute.manager [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Traceback (most recent call last): [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] yield [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] disk_info = self.driver.migrate_disk_and_power_off( [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] disk_key = device.key [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] AttributeError: 'NoneType' object has no attribute 'key' [ 1086.067169] env[61995]: ERROR nova.compute.manager [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] [ 1086.094071] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795262, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.224759} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.094350] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.095177] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e58ae1a-31d2-4d86-a5cc-796d93c7557b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.133238] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.133995] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdb7a003-c341-463a-8d80-2714361a7416 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.162607] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1086.162607] env[61995]: value = "task-795263" [ 1086.162607] env[61995]: _type = "Task" [ 1086.162607] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.171400] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795263, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.302780] env[61995]: DEBUG oslo_concurrency.lockutils [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "83369251-b00f-4595-bf98-28a3ec84f037" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.303070] env[61995]: DEBUG oslo_concurrency.lockutils [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.303317] env[61995]: INFO nova.compute.manager [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Attaching volume ef951e09-0383-4bd5-891c-6e7ead61b641 to /dev/sdb [ 1086.335744] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c46deb-5175-4a03-9de4-9f23fe951b78 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.343248] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1c9bdb-1bb5-4f38-a550-903a2239bd80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.357684] env[61995]: DEBUG nova.virt.block_device [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating existing volume attachment record: d463c72b-6c6d-4926-bace-ea484a3e075a {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1086.462558] env[61995]: DEBUG oslo_concurrency.lockutils [req-5cb9613e-8355-4327-8a18-e24af490ac50 req-469bb3b3-47b6-4f98-87da-fc94e227a045 service nova] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.472012] env[61995]: DEBUG nova.compute.manager [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.473174] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc2d88c-1231-49ba-984b-60d0f3aa8af7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.491082] env[61995]: DEBUG oslo_concurrency.lockutils [req-a0b933c9-aa0e-43fa-8058-bb60d2c8d3a8 req-0e510e37-96a0-4582-b51a-407d25be4b3f service nova] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.571443] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "ed952baf-f1ef-4ec0-8a15-928287190456" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.571784] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "ed952baf-f1ef-4ec0-8a15-928287190456" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.572026] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "ed952baf-f1ef-4ec0-8a15-928287190456-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.572225] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "ed952baf-f1ef-4ec0-8a15-928287190456-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.572394] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "ed952baf-f1ef-4ec0-8a15-928287190456-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.577685] env[61995]: INFO nova.compute.manager [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Terminating instance [ 1086.579821] env[61995]: DEBUG nova.compute.manager [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1086.580028] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1086.580849] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7be3361-11e6-4c67-8010-a9c6c0e933f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.588369] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1086.588634] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57837d3d-c0da-44fb-8af5-01b9b48c13e6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.590915] env[61995]: INFO nova.compute.manager [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Swapping old allocation on dict_keys(['5c086f4d-bc91-4e49-9831-bed8df133c15']) held by migration 8b9e9144-718f-46d0-9851-91e5d87fd2dc for instance [ 1086.598171] env[61995]: DEBUG oslo_vmware.api [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1086.598171] env[61995]: value = "task-795265" [ 1086.598171] env[61995]: _type = "Task" [ 1086.598171] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.605884] env[61995]: DEBUG oslo_vmware.api [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795265, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.618777] env[61995]: DEBUG nova.scheduler.client.report [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Overwriting current allocation {'allocations': {'5c086f4d-bc91-4e49-9831-bed8df133c15': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 135}}, 'project_id': '83064cbae17c429d8d084837635486da', 'user_id': '3efdbd0f8e0340dc8e4bfbf672d62f0e', 'consumer_generation': 1} on consumer ac81c31d-d310-42c7-a6b4-166f3e82cf56 {{(pid=61995) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1086.672522] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795263, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.984224] env[61995]: INFO nova.compute.manager [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] instance snapshotting [ 1086.987089] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e54de1-63be-4203-ae29-c52dca06529a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.005944] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b634dc-93cc-40ec-89e6-fcc864cdef6b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.108928] env[61995]: DEBUG oslo_vmware.api [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795265, 'name': PowerOffVM_Task, 'duration_secs': 0.26583} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.109210] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1087.109383] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1087.109636] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-530658cc-de57-40ee-a4e6-ac7032ec17e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.173863] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795263, 'name': ReconfigVM_Task, 'duration_secs': 0.545223} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.174304] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfigured VM instance instance-00000067 to attach disk [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.174842] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-057c6590-4cb7-4c0b-8968-b4b5cd72ec83 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.181942] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1087.181942] env[61995]: value = "task-795269" [ 1087.181942] env[61995]: _type = "Task" [ 1087.181942] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.183431] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1087.183641] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1087.183826] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleting the datastore file [datastore1] ed952baf-f1ef-4ec0-8a15-928287190456 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.186876] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69e85fea-a2fb-4c11-b734-07720adee7f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.194646] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795269, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.195893] env[61995]: DEBUG oslo_vmware.api [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for the task: (returnval){ [ 1087.195893] env[61995]: value = "task-795270" [ 1087.195893] env[61995]: _type = "Task" [ 1087.195893] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.207175] env[61995]: DEBUG oslo_vmware.api [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795270, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.517902] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.518291] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.518521] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.518715] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.518897] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.521376] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1087.521972] env[61995]: INFO nova.compute.manager [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Terminating instance [ 1087.523314] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-673c6782-fa81-4a49-950e-657ed21defda {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.526140] env[61995]: DEBUG nova.compute.manager [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1087.526418] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1087.526635] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0944db1a-d436-4e91-8767-b82830a61dd4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.533414] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1087.533414] env[61995]: value = "task-795271" [ 1087.533414] env[61995]: _type = "Task" [ 1087.533414] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.534631] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1087.534631] env[61995]: value = "task-795272" [ 1087.534631] env[61995]: _type = "Task" [ 1087.534631] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.544627] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795271, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.547971] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1087.548223] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.548430] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185404', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'name': 'volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac81c31d-d310-42c7-a6b4-166f3e82cf56', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'serial': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.549147] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5dc679-1eec-4900-aef0-8070c7cc41b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.566483] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3880fe-c0ee-4326-988e-7bba81b79601 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.572803] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c52816e-6d84-4adf-acdb-4265648d0a71 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.589809] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14c087e-b158-456d-ba8b-87d321dfc6f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.603990] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] The volume has not been displaced from its original location: [datastore2] volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7/volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1087.609163] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1087.609479] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68a9b2b5-969f-4866-848c-6e25398b4516 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.626426] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1087.626426] env[61995]: value = "task-795273" [ 1087.626426] env[61995]: _type = "Task" [ 1087.626426] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.638132] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795273, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.698912] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795269, 'name': Rename_Task, 'duration_secs': 0.164226} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.701989] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1087.702581] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c55331c9-084f-4556-a6c6-e36013867f07 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.710528] env[61995]: DEBUG oslo_vmware.api [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Task: {'id': task-795270, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138961} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.711887] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1087.712121] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1087.712303] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1087.712486] env[61995]: INFO nova.compute.manager [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1087.712740] env[61995]: DEBUG oslo.service.loopingcall [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1087.713009] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1087.713009] env[61995]: value = "task-795274" [ 1087.713009] env[61995]: _type = "Task" [ 1087.713009] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.713228] env[61995]: DEBUG nova.compute.manager [-] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1087.713302] env[61995]: DEBUG nova.network.neutron [-] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1087.723550] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795274, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.999272] env[61995]: DEBUG nova.compute.manager [req-f4a3e68a-48a7-4a75-935a-1cba8e08d9c5 req-e42f13b8-9368-43e9-bf49-1badb1590a5f service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Received event network-vif-deleted-8df0df4d-9ab4-4900-a107-027ae7ffe387 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.999590] env[61995]: INFO nova.compute.manager [req-f4a3e68a-48a7-4a75-935a-1cba8e08d9c5 req-e42f13b8-9368-43e9-bf49-1badb1590a5f service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Neutron deleted interface 8df0df4d-9ab4-4900-a107-027ae7ffe387; detaching it from the instance and deleting it from the info cache [ 1087.999794] env[61995]: DEBUG nova.network.neutron [req-f4a3e68a-48a7-4a75-935a-1cba8e08d9c5 req-e42f13b8-9368-43e9-bf49-1badb1590a5f service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.043721] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795271, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.136306] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-c3b2e000-cfa8-4b5b-8763-982c6b768621-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.136564] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-c3b2e000-cfa8-4b5b-8763-982c6b768621-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.136932] env[61995]: DEBUG nova.objects.instance [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'flavor' on Instance uuid c3b2e000-cfa8-4b5b-8763-982c6b768621 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.138285] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795273, 'name': ReconfigVM_Task, 'duration_secs': 0.223859} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.138544] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.144366] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a932539-6e9d-4ad9-b7a6-7f76b4692524 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.161574] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1088.161574] env[61995]: value = "task-795275" [ 1088.161574] env[61995]: _type = "Task" [ 1088.161574] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.170937] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.195725] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.196149] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.226963] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795274, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.385020] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4abec9a-80c9-4712-9f34-c5b2f53bff8a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.391734] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea6080a-5033-485c-b372-ffebed20916d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.423545] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c82b91-f5f0-4994-8a04-378081acb260 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.430727] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d3d72e-9d49-48d8-939e-06076d7333a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.443854] env[61995]: DEBUG nova.compute.provider_tree [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.479992] env[61995]: DEBUG nova.network.neutron [-] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.502886] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6873911-1530-4f50-8c0a-d7cd4366bccc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.511897] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ebf87b-c1a0-4c83-a3af-b082f0b769a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.542548] env[61995]: DEBUG nova.compute.manager [req-f4a3e68a-48a7-4a75-935a-1cba8e08d9c5 req-e42f13b8-9368-43e9-bf49-1badb1590a5f service nova] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Detach interface failed, port_id=8df0df4d-9ab4-4900-a107-027ae7ffe387, reason: Instance ed952baf-f1ef-4ec0-8a15-928287190456 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1088.549712] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795271, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.671263] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795275, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.728663] env[61995]: DEBUG oslo_vmware.api [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795274, 'name': PowerOnVM_Task, 'duration_secs': 0.627943} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.729106] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1088.729505] env[61995]: INFO nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Took 9.48 seconds to spawn the instance on the hypervisor. [ 1088.729837] env[61995]: DEBUG nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.730694] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6df316-8b31-4368-ab79-05f3958acf2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.751493] env[61995]: DEBUG nova.objects.instance [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'pci_requests' on Instance uuid c3b2e000-cfa8-4b5b-8763-982c6b768621 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.839720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.947332] env[61995]: DEBUG nova.scheduler.client.report [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.982796] env[61995]: INFO nova.compute.manager [-] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Took 1.27 seconds to deallocate network for instance. [ 1089.052663] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795271, 'name': CreateSnapshot_Task, 'duration_secs': 1.363194} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.054037] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1089.054037] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d02987-eae7-48a0-b44d-e6ecf257f864 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.172298] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795275, 'name': ReconfigVM_Task, 'duration_secs': 0.75231} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.172606] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185404', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'name': 'volume-bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac81c31d-d310-42c7-a6b4-166f3e82cf56', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7', 'serial': 'bbdbeaec-455e-4027-8c73-ed438ff9dde7'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1089.173333] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1089.173645] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31bb961-4596-46c3-a7ad-857e6c034793 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.180219] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1089.180460] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c738007-18f6-4f7c-82ab-1c3e0babd57a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.246734] env[61995]: INFO nova.compute.manager [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Took 18.65 seconds to build instance. [ 1089.249755] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1089.250015] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1089.250253] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleting the datastore file [datastore2] ac81c31d-d310-42c7-a6b4-166f3e82cf56 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.250490] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd9c82dc-4f5a-4e7d-9d2f-284a043b932a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.253261] env[61995]: DEBUG nova.objects.base [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1089.253513] env[61995]: DEBUG nova.network.neutron [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1089.256898] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1089.256898] env[61995]: value = "task-795278" [ 1089.256898] env[61995]: _type = "Task" [ 1089.256898] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.265130] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795278, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.330764] env[61995]: DEBUG nova.policy [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1089.454344] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.256s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.454344] env[61995]: INFO nova.compute.manager [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Successfully reverted task state from resize_migrating on failure for instance. [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server [None req-3e3df7eb-a526-4baa-8496-d525a5549655 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server raise self.value [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1089.469645] env[61995]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1089.471177] env[61995]: ERROR oslo_messaging.rpc.server [ 1089.491348] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.491634] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.491899] env[61995]: DEBUG nova.objects.instance [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lazy-loading 'resources' on Instance uuid ed952baf-f1ef-4ec0-8a15-928287190456 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.508497] env[61995]: DEBUG nova.compute.manager [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Received event network-changed-06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1089.508761] env[61995]: DEBUG nova.compute.manager [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Refreshing instance network info cache due to event network-changed-06334396-e435-480b-ab0c-0da719e15bf1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1089.508910] env[61995]: DEBUG oslo_concurrency.lockutils [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] Acquiring lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.509073] env[61995]: DEBUG oslo_concurrency.lockutils [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] Acquired lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.509279] env[61995]: DEBUG nova.network.neutron [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Refreshing network info cache for port 06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1089.571637] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1089.572626] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8c1f3eb2-547e-4b59-9b61-2464d7540608 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.582536] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1089.582536] env[61995]: value = "task-795279" [ 1089.582536] env[61995]: _type = "Task" [ 1089.582536] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.595441] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795279, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.748958] env[61995]: DEBUG oslo_concurrency.lockutils [None req-32dad923-21dc-4272-943c-6fb2d1cb1e8e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.159s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.768316] env[61995]: DEBUG oslo_vmware.api [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795278, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087646} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.768316] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.768538] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1089.768691] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1089.768843] env[61995]: INFO nova.compute.manager [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 2.24 seconds to destroy the instance on the hypervisor. [ 1089.769992] env[61995]: DEBUG oslo.service.loopingcall [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.769992] env[61995]: DEBUG nova.compute.manager [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1089.769992] env[61995]: DEBUG nova.network.neutron [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1090.096118] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795279, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.208462] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a819ec6-af26-4f45-83e1-d038d6d0a3f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.216256] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43032ca5-ac61-408b-952d-fbe5409ea5ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.224048] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.224388] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.224645] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.224846] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.225056] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.256014] env[61995]: INFO nova.compute.manager [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Terminating instance [ 1090.258020] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6c81a1-b05e-4d5b-88b8-fc52b0a79a22 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.261690] env[61995]: DEBUG nova.compute.manager [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.261763] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1090.265706] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589a7e71-4589-473b-8045-826ef57b29ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.268131] env[61995]: DEBUG nova.compute.manager [req-acb4edfe-1101-40ca-a0e5-1551c157b88b req-35a196db-8f08-4604-93a2-dcaa90fa4cb3 service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Received event network-vif-deleted-7c06dfad-348a-49b5-8d8d-d606f4b559c9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1090.268380] env[61995]: INFO nova.compute.manager [req-acb4edfe-1101-40ca-a0e5-1551c157b88b req-35a196db-8f08-4604-93a2-dcaa90fa4cb3 service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Neutron deleted interface 7c06dfad-348a-49b5-8d8d-d606f4b559c9; detaching it from the instance and deleting it from the info cache [ 1090.268625] env[61995]: DEBUG nova.network.neutron [req-acb4edfe-1101-40ca-a0e5-1551c157b88b req-35a196db-8f08-4604-93a2-dcaa90fa4cb3 service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.274501] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1958f3e-3888-4c4f-9bc0-6ac82e35c3dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.281608] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1090.282221] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5cc3c43e-7f50-4bf6-b71e-700d377080b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.294899] env[61995]: DEBUG nova.compute.provider_tree [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.297574] env[61995]: DEBUG oslo_vmware.api [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1090.297574] env[61995]: value = "task-795280" [ 1090.297574] env[61995]: _type = "Task" [ 1090.297574] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.301818] env[61995]: DEBUG nova.network.neutron [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updated VIF entry in instance network info cache for port 06334396-e435-480b-ab0c-0da719e15bf1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1090.302264] env[61995]: DEBUG nova.network.neutron [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.309793] env[61995]: DEBUG oslo_vmware.api [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795280, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.593117] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795279, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.756781] env[61995]: DEBUG nova.network.neutron [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.774762] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-907afb4b-3ba6-485a-a6b1-a6597fa1e4bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.788064] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b71a460-615d-47f6-a329-5831764194ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.805019] env[61995]: DEBUG nova.scheduler.client.report [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1090.806475] env[61995]: DEBUG oslo_concurrency.lockutils [req-3ee7d53f-facf-4302-a797-e6c5fee41106 req-067baf03-9664-4ece-b49b-8d53dd8faa42 service nova] Releasing lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.818329] env[61995]: DEBUG oslo_vmware.api [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795280, 'name': PowerOffVM_Task, 'duration_secs': 0.202803} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.818975] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1090.819520] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1090.819996] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7dcbe166-b43d-48ae-8022-b1f4458a5e0c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.838667] env[61995]: DEBUG nova.compute.manager [req-acb4edfe-1101-40ca-a0e5-1551c157b88b req-35a196db-8f08-4604-93a2-dcaa90fa4cb3 service nova] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Detach interface failed, port_id=7c06dfad-348a-49b5-8d8d-d606f4b559c9, reason: Instance ac81c31d-d310-42c7-a6b4-166f3e82cf56 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1090.846564] env[61995]: DEBUG nova.network.neutron [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Successfully updated port: 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.888774] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1090.889154] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1090.889457] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleting the datastore file [datastore2] 410eac71-20cd-4a6d-9b78-e11fa72d74e7 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.890071] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fc196cd-cb1b-445e-bfd2-4469a6e9f178 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.900048] env[61995]: DEBUG oslo_vmware.api [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1090.900048] env[61995]: value = "task-795282" [ 1090.900048] env[61995]: _type = "Task" [ 1090.900048] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.905880] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1090.906160] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185418', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'name': 'volume-ef951e09-0383-4bd5-891c-6e7ead61b641', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '83369251-b00f-4595-bf98-28a3ec84f037', 'attached_at': '', 'detached_at': '', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'serial': 'ef951e09-0383-4bd5-891c-6e7ead61b641'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1090.906939] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80985e10-27d7-490a-93c4-6037333400d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.913178] env[61995]: DEBUG oslo_vmware.api [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.925927] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daca5bf6-c44d-4c49-b00e-d89e50e794f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.650870] env[61995]: INFO nova.compute.manager [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 1.88 seconds to deallocate network for instance. [ 1091.651588] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.160s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.654226] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.654481] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.654669] env[61995]: DEBUG nova.network.neutron [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1091.663334] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-ef951e09-0383-4bd5-891c-6e7ead61b641/volume-ef951e09-0383-4bd5-891c-6e7ead61b641.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1091.671763] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3c78b94-6cd1-49df-b76d-143e5a21cddb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.686402] env[61995]: INFO nova.scheduler.client.report [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Deleted allocations for instance ed952baf-f1ef-4ec0-8a15-928287190456 [ 1091.695967] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795279, 'name': CloneVM_Task, 'duration_secs': 1.494305} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.696452] env[61995]: DEBUG oslo_vmware.api [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16358} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.697652] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Created linked-clone VM from snapshot [ 1091.701018] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.701018] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.701018] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.701018] env[61995]: INFO nova.compute.manager [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Took 1.44 seconds to destroy the instance on the hypervisor. [ 1091.701018] env[61995]: DEBUG oslo.service.loopingcall [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.701018] env[61995]: DEBUG oslo_vmware.api [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1091.701018] env[61995]: value = "task-795283" [ 1091.701018] env[61995]: _type = "Task" [ 1091.701018] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.701018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64a822d-0c24-4e1f-882e-b5d078782224 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.702038] env[61995]: DEBUG nova.compute.manager [-] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.702118] env[61995]: DEBUG nova.network.neutron [-] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1091.712050] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Uploading image 258f374f-e693-46c8-87e6-9565d80e03ff {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1091.717835] env[61995]: DEBUG oslo_vmware.api [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.738412] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1091.738412] env[61995]: value = "vm-185420" [ 1091.738412] env[61995]: _type = "VirtualMachine" [ 1091.738412] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1091.738412] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ac0124eb-b702-42af-8282-fb0f205c521e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.747173] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease: (returnval){ [ 1091.747173] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f6fffd-c2e8-ae33-3b39-61f474cee9f4" [ 1091.747173] env[61995]: _type = "HttpNfcLease" [ 1091.747173] env[61995]: } obtained for exporting VM: (result){ [ 1091.747173] env[61995]: value = "vm-185420" [ 1091.747173] env[61995]: _type = "VirtualMachine" [ 1091.747173] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1091.747416] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the lease: (returnval){ [ 1091.747416] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f6fffd-c2e8-ae33-3b39-61f474cee9f4" [ 1091.747416] env[61995]: _type = "HttpNfcLease" [ 1091.747416] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1091.753757] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1091.753757] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f6fffd-c2e8-ae33-3b39-61f474cee9f4" [ 1091.753757] env[61995]: _type = "HttpNfcLease" [ 1091.753757] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1092.202375] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9ab1efd6-0281-4cda-aa91-e7a76e53e2cc tempest-ServersTestJSON-887898215 tempest-ServersTestJSON-887898215-project-member] Lock "ed952baf-f1ef-4ec0-8a15-928287190456" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.630s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.215137] env[61995]: DEBUG oslo_vmware.api [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.225439] env[61995]: WARNING nova.network.neutron [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 1092.230464] env[61995]: INFO nova.compute.manager [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 0.56 seconds to detach 1 volumes for instance. [ 1092.233043] env[61995]: DEBUG nova.compute.manager [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Deleting volume: bbdbeaec-455e-4027-8c73-ed438ff9dde7 {{(pid=61995) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1092.255592] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1092.255592] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f6fffd-c2e8-ae33-3b39-61f474cee9f4" [ 1092.255592] env[61995]: _type = "HttpNfcLease" [ 1092.255592] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1092.255905] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1092.255905] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f6fffd-c2e8-ae33-3b39-61f474cee9f4" [ 1092.255905] env[61995]: _type = "HttpNfcLease" [ 1092.255905] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1092.256689] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7840cc01-31ea-4c15-b749-0629dd948c34 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.264804] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfc54-af1d-3a43-3f66-d19c7f24a799/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1092.264993] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfc54-af1d-3a43-3f66-d19c7f24a799/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1092.339986] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.340273] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.340496] env[61995]: DEBUG nova.objects.instance [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'resources' on Instance uuid ac81c31d-d310-42c7-a6b4-166f3e82cf56 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.344926] env[61995]: DEBUG nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-vif-plugged-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.345557] env[61995]: DEBUG oslo_concurrency.lockutils [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] Acquiring lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.346794] env[61995]: DEBUG oslo_concurrency.lockutils [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.346794] env[61995]: DEBUG oslo_concurrency.lockutils [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.346794] env[61995]: DEBUG nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] No waiting events found dispatching network-vif-plugged-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1092.346794] env[61995]: WARNING nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received unexpected event network-vif-plugged-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 for instance with vm_state active and task_state None. [ 1092.346794] env[61995]: DEBUG nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-changed-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.346794] env[61995]: DEBUG nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing instance network info cache due to event network-changed-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1092.346794] env[61995]: DEBUG oslo_concurrency.lockutils [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.399982] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a023a2c6-aad2-4082-9cb6-588a5c81f761 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.438117] env[61995]: DEBUG nova.network.neutron [-] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.537021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcad0f6-25d1-423d-b8b0-db486783df66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.544543] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551c8669-8747-4df2-8e0f-74bcffc76402 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.576730] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e5c2f5-120b-4d93-9bc3-dcfcc238bb0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.585237] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21730218-180c-4850-99f4-683f95d9a0a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.600853] env[61995]: DEBUG nova.compute.provider_tree [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.624140] env[61995]: DEBUG nova.network.neutron [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "address": "fa:16:3e:5b:d9:e3", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b27da3c-fc", "ovs_interfaceid": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.714551] env[61995]: DEBUG oslo_vmware.api [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795283, 'name': ReconfigVM_Task, 'duration_secs': 0.69436} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.714913] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-ef951e09-0383-4bd5-891c-6e7ead61b641/volume-ef951e09-0383-4bd5-891c-6e7ead61b641.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.719553] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c15ca723-0958-4304-ba3d-345a407d1a29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.735688] env[61995]: DEBUG oslo_vmware.api [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1092.735688] env[61995]: value = "task-795286" [ 1092.735688] env[61995]: _type = "Task" [ 1092.735688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.744818] env[61995]: DEBUG oslo_vmware.api [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795286, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.940609] env[61995]: INFO nova.compute.manager [-] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Took 1.24 seconds to deallocate network for instance. [ 1093.104993] env[61995]: DEBUG nova.scheduler.client.report [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.127265] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.128071] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.128572] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.129159] env[61995]: DEBUG oslo_concurrency.lockutils [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.129454] env[61995]: DEBUG nova.network.neutron [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing network info cache for port 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1093.131818] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bac75b-5ef9-4d43-92c5-00cf060a0bc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.152216] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1093.153419] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1093.153721] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1093.154024] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1093.154272] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1093.154466] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1093.154981] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1093.155323] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1093.155972] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1093.156293] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1093.157299] env[61995]: DEBUG nova.virt.hardware [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.163231] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Reconfiguring VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1093.164017] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db9d4e87-27c2-4c3a-9763-185394d51482 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.183516] env[61995]: DEBUG oslo_vmware.api [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1093.183516] env[61995]: value = "task-795287" [ 1093.183516] env[61995]: _type = "Task" [ 1093.183516] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.191815] env[61995]: DEBUG oslo_vmware.api [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795287, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.246535] env[61995]: DEBUG oslo_vmware.api [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795286, 'name': ReconfigVM_Task, 'duration_secs': 0.183011} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.247033] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185418', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'name': 'volume-ef951e09-0383-4bd5-891c-6e7ead61b641', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '83369251-b00f-4595-bf98-28a3ec84f037', 'attached_at': '', 'detached_at': '', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'serial': 'ef951e09-0383-4bd5-891c-6e7ead61b641'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1093.448470] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.612627] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.272s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.615117] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.167s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.615422] env[61995]: DEBUG nova.objects.instance [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'resources' on Instance uuid 410eac71-20cd-4a6d-9b78-e11fa72d74e7 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.661667] env[61995]: INFO nova.scheduler.client.report [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted allocations for instance ac81c31d-d310-42c7-a6b4-166f3e82cf56 [ 1093.695473] env[61995]: DEBUG oslo_vmware.api [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795287, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.878735] env[61995]: DEBUG nova.network.neutron [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updated VIF entry in instance network info cache for port 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.878735] env[61995]: DEBUG nova.network.neutron [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "address": "fa:16:3e:5b:d9:e3", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b27da3c-fc", "ovs_interfaceid": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.176421] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed8d217d-5958-4066-b9d3-0254194dd5e8 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.658s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.177809] env[61995]: DEBUG oslo_concurrency.lockutils [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.338s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.178243] env[61995]: DEBUG oslo_concurrency.lockutils [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.179160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.180616] env[61995]: DEBUG oslo_concurrency.lockutils [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.186215] env[61995]: INFO nova.compute.manager [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Terminating instance [ 1094.190639] env[61995]: DEBUG nova.compute.manager [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1094.190964] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0347851-fa5e-4ecf-a39b-19aedfcff10a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.198735] env[61995]: DEBUG oslo_vmware.api [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795287, 'name': ReconfigVM_Task, 'duration_secs': 0.611541} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.200346] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.200448] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Reconfigured VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1094.209142] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f560a88-4fac-40c8-93aa-f9ae8ebfd3bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.244677] env[61995]: WARNING nova.virt.vmwareapi.driver [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance ac81c31d-d310-42c7-a6b4-166f3e82cf56 could not be found. [ 1094.244910] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1094.247927] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5bccd9e8-8295-406f-9cf1-18d5216e8d12 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.256919] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d1f4cb-8a60-4bde-b5c7-739fb17d3a00 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.290611] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ac81c31d-d310-42c7-a6b4-166f3e82cf56 could not be found. [ 1094.290829] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1094.291747] env[61995]: INFO nova.compute.manager [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1094.291747] env[61995]: DEBUG oslo.service.loopingcall [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.294264] env[61995]: DEBUG nova.compute.manager [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1094.294851] env[61995]: DEBUG nova.network.neutron [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1094.302346] env[61995]: DEBUG nova.objects.instance [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'flavor' on Instance uuid 83369251-b00f-4595-bf98-28a3ec84f037 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.381621] env[61995]: DEBUG oslo_concurrency.lockutils [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.381960] env[61995]: DEBUG nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Received event network-vif-deleted-505624a2-00f0-4314-b814-28c5bc7ac5d9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.382186] env[61995]: INFO nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Neutron deleted interface 505624a2-00f0-4314-b814-28c5bc7ac5d9; detaching it from the instance and deleting it from the info cache [ 1094.382370] env[61995]: DEBUG nova.network.neutron [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.394805] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5b9aae-4451-4382-9178-095dbf3535af {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.403532] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460897ce-5d71-41f0-a550-8850dc83f196 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.436518] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27677b1d-5c83-4f95-a468-5fc279a73305 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.444839] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da79a1e9-cf47-492b-bb73-b4e58cddee3f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.459623] env[61995]: DEBUG nova.compute.provider_tree [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.723142] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f95b5737-8b5d-44cb-9732-aa87f8af2801 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-c3b2e000-cfa8-4b5b-8763-982c6b768621-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.586s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.807510] env[61995]: DEBUG oslo_concurrency.lockutils [None req-717ecbd5-5458-4c48-a45d-12b241f6e3a7 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.504s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.820351] env[61995]: DEBUG nova.network.neutron [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.884861] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-274d469b-482e-477d-983b-f079b87bdeab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.894344] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84aec0b3-70ad-4b6d-a4b1-5a49cf9e7db2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.927820] env[61995]: DEBUG nova.compute.manager [req-45d6407e-c463-4b00-b52f-8b4a587013bc req-c1c05fa1-a0bb-4ced-8270-0a623bff795d service nova] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Detach interface failed, port_id=505624a2-00f0-4314-b814-28c5bc7ac5d9, reason: Instance 410eac71-20cd-4a6d-9b78-e11fa72d74e7 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1094.963292] env[61995]: DEBUG nova.scheduler.client.report [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.323024] env[61995]: INFO nova.compute.manager [-] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 1.03 seconds to deallocate network for instance. [ 1095.337713] env[61995]: WARNING nova.volume.cinder [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Attachment c88cf9d4-9142-47b6-a301-3de841f75b6f does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = c88cf9d4-9142-47b6-a301-3de841f75b6f. (HTTP 404) (Request-ID: req-b28cc702-5ce9-49a5-94b6-afee0b729aad) [ 1095.338209] env[61995]: INFO nova.compute.manager [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Took 0.01 seconds to detach 1 volumes for instance. [ 1095.340985] env[61995]: DEBUG nova.compute.manager [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Deleting volume: bbdbeaec-455e-4027-8c73-ed438ff9dde7 {{(pid=61995) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1095.356384] env[61995]: WARNING nova.compute.manager [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Failed to delete volume: bbdbeaec-455e-4027-8c73-ed438ff9dde7 due to Volume bbdbeaec-455e-4027-8c73-ed438ff9dde7 could not be found.: nova.exception.VolumeNotFound: Volume bbdbeaec-455e-4027-8c73-ed438ff9dde7 could not be found. [ 1095.468788] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.495538] env[61995]: INFO nova.scheduler.client.report [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted allocations for instance 410eac71-20cd-4a6d-9b78-e11fa72d74e7 [ 1095.594065] env[61995]: DEBUG nova.compute.manager [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1095.861759] env[61995]: INFO nova.compute.manager [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance disappeared during terminate [ 1095.862160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-08d8ca7c-0a00-4c48-8218-e08736339ed0 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "ac81c31d-d310-42c7-a6b4-166f3e82cf56" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.684s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.003915] env[61995]: DEBUG oslo_concurrency.lockutils [None req-83c3b7a0-7427-47a6-8976-27cf6df2c5e2 tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "410eac71-20cd-4a6d-9b78-e11fa72d74e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.779s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.111983] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.112294] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.355146] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-c3b2e000-cfa8-4b5b-8763-982c6b768621-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.355431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-c3b2e000-cfa8-4b5b-8763-982c6b768621-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.618487] env[61995]: INFO nova.compute.claims [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1096.859139] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.859368] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.860359] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2b6836-d9f3-4402-ba78-f58e80254c1f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.878908] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7915b32-58e5-426f-9e20-43806acbc9c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.906421] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Reconfiguring VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1096.907054] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d68f287e-b3e9-4343-8b77-a824e4fb5c3d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.926608] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1096.926608] env[61995]: value = "task-795289" [ 1096.926608] env[61995]: _type = "Task" [ 1096.926608] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.934505] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.124934] env[61995]: INFO nova.compute.resource_tracker [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating resource usage from migration 20ca0923-85a1-4eb9-89df-078708479759 [ 1097.285328] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b3e4f5-7225-4fdb-93a0-6805273e0300 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.293412] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a670b8-d210-4e1a-97c0-538aa4887f10 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.324527] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54afb5f-7b81-4df1-9125-a13dfeff4e9c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.332186] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a35cd49-78be-4663-ab0d-a49536764d7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.345013] env[61995]: DEBUG nova.compute.provider_tree [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.437754] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.848866] env[61995]: DEBUG nova.scheduler.client.report [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.937780] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.353603] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.241s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.353875] env[61995]: INFO nova.compute.manager [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Migrating [ 1098.442657] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.869862] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.870095] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.870293] env[61995]: DEBUG nova.network.neutron [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1098.940485] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.442324] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.683324] env[61995]: DEBUG nova.network.neutron [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [{"id": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "address": "fa:16:3e:ae:d5:b8", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc97b92c2-09", "ovs_interfaceid": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.934980] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "b40b1866-ced3-40ef-9ab7-30d58b75f288" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.935266] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.935485] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "b40b1866-ced3-40ef-9ab7-30d58b75f288-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.935679] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.935857] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.941089] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.941624] env[61995]: INFO nova.compute.manager [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Terminating instance [ 1099.943912] env[61995]: DEBUG nova.compute.manager [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1099.944450] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1099.944919] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f4479e-49e5-422d-8d45-ca81c3c17cb9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.952851] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1099.953458] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70003f5f-3ffb-4f54-a9b9-9e13e26fc33d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.964551] env[61995]: DEBUG oslo_vmware.api [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1099.964551] env[61995]: value = "task-795290" [ 1099.964551] env[61995]: _type = "Task" [ 1099.964551] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.971206] env[61995]: DEBUG oslo_vmware.api [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795290, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.187159] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.444145] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.473098] env[61995]: DEBUG oslo_vmware.api [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795290, 'name': PowerOffVM_Task, 'duration_secs': 0.307871} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.473407] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1100.473578] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1100.473834] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de2231eb-c2c7-477e-9b4c-64f862243b7a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.548354] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1100.548694] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1100.548927] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleting the datastore file [datastore2] b40b1866-ced3-40ef-9ab7-30d58b75f288 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1100.549356] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e43bbb9-0709-48e7-9b55-c959c676081d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.555495] env[61995]: DEBUG oslo_vmware.api [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for the task: (returnval){ [ 1100.555495] env[61995]: value = "task-795292" [ 1100.555495] env[61995]: _type = "Task" [ 1100.555495] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.564017] env[61995]: DEBUG oslo_vmware.api [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.945294] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.068843] env[61995]: DEBUG oslo_vmware.api [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Task: {'id': task-795292, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.393197} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.069304] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1101.069560] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1101.069791] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1101.069990] env[61995]: INFO nova.compute.manager [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1101.070320] env[61995]: DEBUG oslo.service.loopingcall [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1101.070657] env[61995]: DEBUG nova.compute.manager [-] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1101.070787] env[61995]: DEBUG nova.network.neutron [-] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1101.344041] env[61995]: DEBUG nova.compute.manager [req-c4c5e5a9-f35d-492b-aa0c-f8eab6709ec5 req-cc632211-2b03-4124-84a8-a9357931f439 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Received event network-vif-deleted-ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1101.344333] env[61995]: INFO nova.compute.manager [req-c4c5e5a9-f35d-492b-aa0c-f8eab6709ec5 req-cc632211-2b03-4124-84a8-a9357931f439 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Neutron deleted interface ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f; detaching it from the instance and deleting it from the info cache [ 1101.344604] env[61995]: DEBUG nova.network.neutron [req-c4c5e5a9-f35d-492b-aa0c-f8eab6709ec5 req-cc632211-2b03-4124-84a8-a9357931f439 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.444864] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.704995] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372bbeb8-2642-4aef-a239-2a87673c8dd4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.727014] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance '83369251-b00f-4595-bf98-28a3ec84f037' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1101.814808] env[61995]: DEBUG nova.network.neutron [-] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.847299] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8973edfa-773e-4cdf-9851-ff3ad7a151ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.856923] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54a9002-3c59-415b-b188-f221080b6e96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.890330] env[61995]: DEBUG nova.compute.manager [req-c4c5e5a9-f35d-492b-aa0c-f8eab6709ec5 req-cc632211-2b03-4124-84a8-a9357931f439 service nova] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Detach interface failed, port_id=ddcee8fc-e0b1-4c5f-8e95-aef5e24f0b0f, reason: Instance b40b1866-ced3-40ef-9ab7-30d58b75f288 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1101.915213] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfc54-af1d-3a43-3f66-d19c7f24a799/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1101.916195] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085498d7-d3d8-42ec-8cb5-8dabca4c04a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.922681] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfc54-af1d-3a43-3f66-d19c7f24a799/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1101.922777] env[61995]: ERROR oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfc54-af1d-3a43-3f66-d19c7f24a799/disk-0.vmdk due to incomplete transfer. [ 1101.922914] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-407927bb-c5a7-4cae-8bdd-1baeec888ece {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.930022] env[61995]: DEBUG oslo_vmware.rw_handles [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfc54-af1d-3a43-3f66-d19c7f24a799/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1101.930022] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Uploaded image 258f374f-e693-46c8-87e6-9565d80e03ff to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1101.932970] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1101.933234] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-11a999ba-3ea5-49a8-b4e0-3711458d6870 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.939857] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1101.939857] env[61995]: value = "task-795293" [ 1101.939857] env[61995]: _type = "Task" [ 1101.939857] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.945949] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.950559] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795293, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.985686] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.985977] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.233880] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1102.234237] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9dd193a-0c4a-47f6-ba85-6b9999a78318 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.241973] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1102.241973] env[61995]: value = "task-795294" [ 1102.241973] env[61995]: _type = "Task" [ 1102.241973] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.250894] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.318064] env[61995]: INFO nova.compute.manager [-] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Took 1.25 seconds to deallocate network for instance. [ 1102.449323] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.452470] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795293, 'name': Destroy_Task, 'duration_secs': 0.344794} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.452560] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Destroyed the VM [ 1102.452821] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1102.453108] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5138e6e2-cbba-466d-835e-64212f3bc65c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.462594] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1102.462594] env[61995]: value = "task-795295" [ 1102.462594] env[61995]: _type = "Task" [ 1102.462594] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.471986] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795295, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.489617] env[61995]: DEBUG nova.compute.utils [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1102.751807] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795294, 'name': PowerOffVM_Task, 'duration_secs': 0.278764} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.752107] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1102.752306] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance '83369251-b00f-4595-bf98-28a3ec84f037' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1102.824830] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.825134] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.825334] env[61995]: DEBUG nova.objects.instance [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lazy-loading 'resources' on Instance uuid b40b1866-ced3-40ef-9ab7-30d58b75f288 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.946757] env[61995]: DEBUG oslo_vmware.api [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795289, 'name': ReconfigVM_Task, 'duration_secs': 5.834649} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.947016] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.947241] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Reconfigured VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1102.971787] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795295, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.992016] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.259296] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.259566] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.259730] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.259924] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.260099] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.260264] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.260548] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.260905] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.260905] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.261065] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.261255] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.266779] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2670599e-e1f6-4677-af7e-d8535c0cd7ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.282110] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1103.282110] env[61995]: value = "task-795296" [ 1103.282110] env[61995]: _type = "Task" [ 1103.282110] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.290500] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795296, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.472495] env[61995]: DEBUG oslo_vmware.api [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795295, 'name': RemoveSnapshot_Task, 'duration_secs': 0.65612} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.472863] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1103.472986] env[61995]: INFO nova.compute.manager [None req-84b0c96e-7748-4a4c-9bf1-e8299ecf5733 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Took 16.49 seconds to snapshot the instance on the hypervisor. [ 1103.514274] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12e1896-a34b-4223-b50a-8d28db41d213 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.522478] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366e7ec3-90f2-4848-ba14-657268148a75 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.557238] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d12e99-f090-468c-8640-e0aa75fee61a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.568356] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e571e31-4f30-4429-9e1d-2dbedadb3c71 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.582945] env[61995]: DEBUG nova.compute.provider_tree [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.604466] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.604703] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.792100] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795296, 'name': ReconfigVM_Task, 'duration_secs': 0.376225} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.792429] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance '83369251-b00f-4595-bf98-28a3ec84f037' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1104.066014] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.066319] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.066556] env[61995]: INFO nova.compute.manager [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Attaching volume ae89e4f5-9640-4273-a471-d85d8a021c3b to /dev/sdb [ 1104.085900] env[61995]: DEBUG nova.scheduler.client.report [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.099473] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7daaf65d-2b9e-4bc5-9cab-aca6edc0dd27 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.106780] env[61995]: DEBUG nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1104.109687] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c63ead-cd6c-431e-93ca-5821e1be8f27 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.123450] env[61995]: DEBUG nova.virt.block_device [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Updating existing volume attachment record: 3a85d3f1-a061-4e38-a2b8-afef77eca337 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1104.256808] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.257038] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.257232] env[61995]: DEBUG nova.network.neutron [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1104.300134] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.300428] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.300593] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.300778] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.300929] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.301092] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.301303] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.301478] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.301652] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.301818] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.301993] env[61995]: DEBUG nova.virt.hardware [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.308149] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1104.308719] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aef257d9-78c8-42c7-8b90-d59ce876cfd2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.325769] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1104.325769] env[61995]: value = "task-795298" [ 1104.325769] env[61995]: _type = "Task" [ 1104.325769] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.336014] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795298, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.591210] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.766s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.626514] env[61995]: INFO nova.scheduler.client.report [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Deleted allocations for instance b40b1866-ced3-40ef-9ab7-30d58b75f288 [ 1104.631869] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.631869] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.633232] env[61995]: INFO nova.compute.claims [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1104.835355] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795298, 'name': ReconfigVM_Task, 'duration_secs': 0.20806} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.835862] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1104.836629] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709eafaf-b8f8-4bd4-941e-9766e27d0776 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.861202] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.861549] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a358b1c-68ea-4ded-924c-bd1e78c2b685 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.880519] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1104.880519] env[61995]: value = "task-795299" [ 1104.880519] env[61995]: _type = "Task" [ 1104.880519] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.888178] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795299, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.118338] env[61995]: INFO nova.network.neutron [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Port 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1105.118338] env[61995]: DEBUG nova.network.neutron [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.145190] env[61995]: DEBUG oslo_concurrency.lockutils [None req-76f40ac9-d7dc-4d56-ae36-60122aaefc2c tempest-ServerActionsTestOtherA-1132834204 tempest-ServerActionsTestOtherA-1132834204-project-member] Lock "b40b1866-ced3-40ef-9ab7-30d58b75f288" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.210s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.212009] env[61995]: DEBUG nova.compute.manager [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1105.212291] env[61995]: DEBUG nova.compute.manager [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing instance network info cache due to event network-changed-7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1105.212511] env[61995]: DEBUG oslo_concurrency.lockutils [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] Acquiring lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.390202] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795299, 'name': ReconfigVM_Task, 'duration_secs': 0.314324} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.390412] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.390774] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance '83369251-b00f-4595-bf98-28a3ec84f037' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1105.619392] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.621743] env[61995]: DEBUG oslo_concurrency.lockutils [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] Acquired lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.622343] env[61995]: DEBUG nova.network.neutron [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Refreshing network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1105.799978] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fccd848-d68b-4b1d-bf2b-b7655a213b0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.807412] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3b84f4-5e72-492a-90c9-8f8c611b6cd4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.836745] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68c4274-fdcd-4157-90c1-8929f0e2a637 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.843814] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666b1907-e4ba-4541-9acf-99abcbd0d1f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.857557] env[61995]: DEBUG nova.compute.provider_tree [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1105.897298] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9754418e-4228-4af8-82e8-6454e7b76e99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.918694] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d7586a-b028-4f77-b4c6-36ea8a3b6d2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.921384] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-50717fa4-956c-4e59-ab07-e6e8ec6f16bf-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.921616] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-50717fa4-956c-4e59-ab07-e6e8ec6f16bf-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.921960] env[61995]: DEBUG nova.objects.instance [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'flavor' on Instance uuid 50717fa4-956c-4e59-ab07-e6e8ec6f16bf {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.945308] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance '83369251-b00f-4595-bf98-28a3ec84f037' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1106.127657] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a959244b-1f34-4d40-a22b-d239f289bd42 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-c3b2e000-cfa8-4b5b-8763-982c6b768621-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.772s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.361137] env[61995]: DEBUG nova.scheduler.client.report [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.401425] env[61995]: DEBUG nova.network.neutron [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updated VIF entry in instance network info cache for port 7efb2852-0c7e-4e86-8996-8cba798b76b2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1106.401791] env[61995]: DEBUG nova.network.neutron [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [{"id": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "address": "fa:16:3e:01:3b:8c", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7efb2852-0c", "ovs_interfaceid": "7efb2852-0c7e-4e86-8996-8cba798b76b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.541489] env[61995]: DEBUG nova.objects.instance [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'pci_requests' on Instance uuid 50717fa4-956c-4e59-ab07-e6e8ec6f16bf {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.555432] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.555689] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.555836] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1106.868972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.237s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.869557] env[61995]: DEBUG nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1106.904728] env[61995]: DEBUG oslo_concurrency.lockutils [req-bbe6c90c-cf5e-4bc5-b7e9-ebd19dd76127 req-a32089fd-8d07-4278-97d6-2243ce779ec2 service nova] Releasing lock "refresh_cache-c3b2e000-cfa8-4b5b-8763-982c6b768621" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.044020] env[61995]: DEBUG nova.objects.base [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Object Instance<50717fa4-956c-4e59-ab07-e6e8ec6f16bf> lazy-loaded attributes: flavor,pci_requests {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1107.044156] env[61995]: DEBUG nova.network.neutron [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1107.119860] env[61995]: DEBUG nova.policy [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c55029927a4e0595bd2e2e0309319d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a660ec6d4d7e4e76827642cf247f53c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1107.265579] env[61995]: DEBUG nova.compute.manager [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1107.265854] env[61995]: DEBUG nova.compute.manager [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing instance network info cache due to event network-changed-3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1107.266376] env[61995]: DEBUG oslo_concurrency.lockutils [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.266561] env[61995]: DEBUG oslo_concurrency.lockutils [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.266738] env[61995]: DEBUG nova.network.neutron [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1107.374458] env[61995]: DEBUG nova.compute.utils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1107.376869] env[61995]: DEBUG nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1107.376869] env[61995]: DEBUG nova.network.neutron [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1107.424129] env[61995]: DEBUG nova.policy [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '671c956f9796411e9b66268420075628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26c7cc2bdcb6462d8154127098e94875', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1107.621823] env[61995]: DEBUG nova.network.neutron [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Port c97b92c2-09c6-42ae-bac7-ad1583a82097 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1107.726624] env[61995]: DEBUG nova.network.neutron [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Successfully created port: e14467f7-216c-4d82-b986-9e488c9b78b6 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1107.881326] env[61995]: DEBUG nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1108.112553] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.112718] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.112869] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1108.164809] env[61995]: DEBUG nova.network.neutron [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updated VIF entry in instance network info cache for port 3cd23361-0e98-4153-b3ea-06ce74356da6. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1108.165405] env[61995]: DEBUG nova.network.neutron [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.656955] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "83369251-b00f-4595-bf98-28a3ec84f037-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.657239] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.657426] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.669134] env[61995]: DEBUG oslo_concurrency.lockutils [req-f04b1be1-0cae-4b10-881b-b870ee788235 req-29c62ffc-61c6-45ed-8993-00be942ba78f service nova] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.671997] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1108.672285] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185421', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'name': 'volume-ae89e4f5-9640-4273-a471-d85d8a021c3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e2014437-fbcd-454a-893f-9ad7c7d461d0', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'serial': 'ae89e4f5-9640-4273-a471-d85d8a021c3b'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1108.673859] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509f31d0-5ea4-4ac6-9dc8-d3821b42d9c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.690873] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39688af2-113c-49fa-bf68-3b0e7dd8b9a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.716572] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] volume-ae89e4f5-9640-4273-a471-d85d8a021c3b/volume-ae89e4f5-9640-4273-a471-d85d8a021c3b.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.716899] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd381708-8d0c-446b-8036-6e5a0bdcde6a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.735728] env[61995]: DEBUG oslo_vmware.api [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1108.735728] env[61995]: value = "task-795301" [ 1108.735728] env[61995]: _type = "Task" [ 1108.735728] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.743691] env[61995]: DEBUG oslo_vmware.api [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795301, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.891915] env[61995]: DEBUG nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1108.920849] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c413c4443998afd57abbc1df189b4bdb',container_format='bare',created_at=2024-09-10T09:02:17Z,direct_url=,disk_format='vmdk',id=258f374f-e693-46c8-87e6-9565d80e03ff,min_disk=1,min_ram=0,name='tempest-test-snap-1036719986',owner='26c7cc2bdcb6462d8154127098e94875',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-09-10T09:02:33Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.921101] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.921296] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.921456] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.921605] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.921756] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.921970] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.922192] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.922321] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.922494] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.922669] env[61995]: DEBUG nova.virt.hardware [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.923528] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed215ae-6cf4-4536-8131-28d9c97802ec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.931737] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf0db45-8167-4c96-9d46-846eee758664 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.217932] env[61995]: DEBUG nova.network.neutron [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Successfully updated port: 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1109.250199] env[61995]: DEBUG oslo_vmware.api [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795301, 'name': ReconfigVM_Task, 'duration_secs': 0.375314} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.251765] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Reconfigured VM instance instance-00000061 to attach disk [datastore2] volume-ae89e4f5-9640-4273-a471-d85d8a021c3b/volume-ae89e4f5-9640-4273-a471-d85d8a021c3b.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1109.258077] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7652e6f-aef6-4ee7-8bfe-eeecea9611f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.270504] env[61995]: DEBUG nova.compute.manager [req-b48be57e-1677-4aa7-b026-8b0cec1ded68 req-67df5c11-6ec3-406d-bb92-fb1ea0042e47 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Received event network-vif-plugged-e14467f7-216c-4d82-b986-9e488c9b78b6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.270768] env[61995]: DEBUG oslo_concurrency.lockutils [req-b48be57e-1677-4aa7-b026-8b0cec1ded68 req-67df5c11-6ec3-406d-bb92-fb1ea0042e47 service nova] Acquiring lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.271354] env[61995]: DEBUG oslo_concurrency.lockutils [req-b48be57e-1677-4aa7-b026-8b0cec1ded68 req-67df5c11-6ec3-406d-bb92-fb1ea0042e47 service nova] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.271663] env[61995]: DEBUG oslo_concurrency.lockutils [req-b48be57e-1677-4aa7-b026-8b0cec1ded68 req-67df5c11-6ec3-406d-bb92-fb1ea0042e47 service nova] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.271771] env[61995]: DEBUG nova.compute.manager [req-b48be57e-1677-4aa7-b026-8b0cec1ded68 req-67df5c11-6ec3-406d-bb92-fb1ea0042e47 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] No waiting events found dispatching network-vif-plugged-e14467f7-216c-4d82-b986-9e488c9b78b6 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1109.271906] env[61995]: WARNING nova.compute.manager [req-b48be57e-1677-4aa7-b026-8b0cec1ded68 req-67df5c11-6ec3-406d-bb92-fb1ea0042e47 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Received unexpected event network-vif-plugged-e14467f7-216c-4d82-b986-9e488c9b78b6 for instance with vm_state building and task_state spawning. [ 1109.278994] env[61995]: DEBUG oslo_vmware.api [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1109.278994] env[61995]: value = "task-795302" [ 1109.278994] env[61995]: _type = "Task" [ 1109.278994] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.288511] env[61995]: DEBUG oslo_vmware.api [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795302, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.296857] env[61995]: DEBUG nova.compute.manager [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-vif-plugged-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.297089] env[61995]: DEBUG oslo_concurrency.lockutils [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] Acquiring lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.297305] env[61995]: DEBUG oslo_concurrency.lockutils [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.297477] env[61995]: DEBUG oslo_concurrency.lockutils [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.297748] env[61995]: DEBUG nova.compute.manager [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] No waiting events found dispatching network-vif-plugged-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1109.297811] env[61995]: WARNING nova.compute.manager [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received unexpected event network-vif-plugged-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 for instance with vm_state active and task_state None. [ 1109.297957] env[61995]: DEBUG nova.compute.manager [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-changed-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.298189] env[61995]: DEBUG nova.compute.manager [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing instance network info cache due to event network-changed-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1109.298478] env[61995]: DEBUG oslo_concurrency.lockutils [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.298657] env[61995]: DEBUG oslo_concurrency.lockutils [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.298844] env[61995]: DEBUG nova.network.neutron [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Refreshing network info cache for port 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1109.434689] env[61995]: DEBUG nova.network.neutron [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Successfully updated port: e14467f7-216c-4d82-b986-9e488c9b78b6 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1109.437106] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updating instance_info_cache with network_info: [{"id": "2327f433-5c63-4d83-8050-6661d9ad60eb", "address": "fa:16:3e:d0:20:ad", "network": {"id": "8a3b36b7-2188-4a28-9316-f5686fad4d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-888514135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f543b47c4254f4ebeca453aea3123c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2327f433-5c", "ovs_interfaceid": "2327f433-5c63-4d83-8050-6661d9ad60eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.693475] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.693680] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.693863] env[61995]: DEBUG nova.network.neutron [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1109.727041] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.789548] env[61995]: DEBUG oslo_vmware.api [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795302, 'name': ReconfigVM_Task, 'duration_secs': 0.143865} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.789858] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185421', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'name': 'volume-ae89e4f5-9640-4273-a471-d85d8a021c3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e2014437-fbcd-454a-893f-9ad7c7d461d0', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'serial': 'ae89e4f5-9640-4273-a471-d85d8a021c3b'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1109.939533] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "refresh_cache-55a1c5d3-7e00-40f6-9b06-f206bf2bf895" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.939796] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "refresh_cache-55a1c5d3-7e00-40f6-9b06-f206bf2bf895" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.940128] env[61995]: DEBUG nova.network.neutron [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1109.941391] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-24b974c0-9567-46a3-96f6-fac1e602f46e" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.941611] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1109.941946] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.942354] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.942508] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.942648] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.942793] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.942934] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.943080] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1109.943229] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.075438] env[61995]: DEBUG nova.network.neutron [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Added VIF to instance network info cache for port 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 1110.075572] env[61995]: DEBUG nova.network.neutron [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "address": "fa:16:3e:5b:d9:e3", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b27da3c-fc", "ovs_interfaceid": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.451045] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.451045] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.451045] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.451045] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1110.451045] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c2979a-fe1f-4faa-ab3e-1093e1c3a371 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.463315] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa34891-6d70-46e2-bfcd-c9617aebfe0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.481511] env[61995]: DEBUG nova.network.neutron [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1110.483958] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c31c64-3618-443a-8c2e-ed28ae7839ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.491714] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e5a22b-d870-4818-ad07-5e9bb67801c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.524990] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179663MB free_disk=66GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1110.525360] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.525427] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.579330] env[61995]: DEBUG oslo_concurrency.lockutils [req-88860d2b-35dd-433f-9764-dec4c0d353f6 req-b6eee9c4-7e3d-4663-8f55-eaf4da879290 service nova] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.582204] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.582411] env[61995]: DEBUG nova.network.neutron [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1110.676818] env[61995]: DEBUG nova.network.neutron [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Updating instance_info_cache with network_info: [{"id": "e14467f7-216c-4d82-b986-9e488c9b78b6", "address": "fa:16:3e:47:40:53", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape14467f7-21", "ovs_interfaceid": "e14467f7-216c-4d82-b986-9e488c9b78b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.813472] env[61995]: DEBUG nova.network.neutron [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [{"id": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "address": "fa:16:3e:ae:d5:b8", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc97b92c2-09", "ovs_interfaceid": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.839049] env[61995]: DEBUG nova.objects.instance [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'flavor' on Instance uuid e2014437-fbcd-454a-893f-9ad7c7d461d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.870320] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "966783cb-b826-4983-9b77-32ab4ba56728" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.870584] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "966783cb-b826-4983-9b77-32ab4ba56728" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.120286] env[61995]: WARNING nova.network.neutron [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 1111.120539] env[61995]: WARNING nova.network.neutron [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] 67eef3c3-44a6-48d3-9548-e77c76e98379 already exists in list: networks containing: ['67eef3c3-44a6-48d3-9548-e77c76e98379']. ignoring it [ 1111.120717] env[61995]: WARNING nova.network.neutron [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 already exists in list: port_ids containing: ['5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1']. ignoring it [ 1111.182696] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "refresh_cache-55a1c5d3-7e00-40f6-9b06-f206bf2bf895" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.183036] env[61995]: DEBUG nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Instance network_info: |[{"id": "e14467f7-216c-4d82-b986-9e488c9b78b6", "address": "fa:16:3e:47:40:53", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape14467f7-21", "ovs_interfaceid": "e14467f7-216c-4d82-b986-9e488c9b78b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1111.183547] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:40:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e14467f7-216c-4d82-b986-9e488c9b78b6', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1111.191920] env[61995]: DEBUG oslo.service.loopingcall [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.192190] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1111.192511] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b19adac-eda1-4407-9c51-886cd21e22e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.217175] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1111.217175] env[61995]: value = "task-795303" [ 1111.217175] env[61995]: _type = "Task" [ 1111.217175] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.225314] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795303, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.290648] env[61995]: DEBUG nova.compute.manager [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Received event network-changed-e14467f7-216c-4d82-b986-9e488c9b78b6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.290728] env[61995]: DEBUG nova.compute.manager [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Refreshing instance network info cache due to event network-changed-e14467f7-216c-4d82-b986-9e488c9b78b6. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1111.291671] env[61995]: DEBUG oslo_concurrency.lockutils [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] Acquiring lock "refresh_cache-55a1c5d3-7e00-40f6-9b06-f206bf2bf895" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.291671] env[61995]: DEBUG oslo_concurrency.lockutils [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] Acquired lock "refresh_cache-55a1c5d3-7e00-40f6-9b06-f206bf2bf895" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.291671] env[61995]: DEBUG nova.network.neutron [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Refreshing network info cache for port e14467f7-216c-4d82-b986-9e488c9b78b6 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1111.315177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.343697] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5c8aae19-a535-4304-a6ec-cc06795d1d4b tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.277s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.374726] env[61995]: DEBUG nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1111.439548] env[61995]: DEBUG nova.network.neutron [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "address": "fa:16:3e:5b:d9:e3", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b27da3c-fc", "ovs_interfaceid": "5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.535577] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Applying migration context for instance 83369251-b00f-4595-bf98-28a3ec84f037 as it has an incoming, in-progress migration 20ca0923-85a1-4eb9-89df-078708479759. Migration status is post-migrating {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1111.537231] env[61995]: INFO nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating resource usage from migration 20ca0923-85a1-4eb9-89df-078708479759 [ 1111.561082] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 24b974c0-9567-46a3-96f6-fac1e602f46e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561082] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 5cb4c1d9-e661-48b4-8600-1b823d63ddbb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561264] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 66c56f7b-1fc7-4aed-8afc-350817e1ca48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561350] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance c3b2e000-cfa8-4b5b-8763-982c6b768621 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561470] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance e2014437-fbcd-454a-893f-9ad7c7d461d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561583] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance a73742da-3d5c-4b71-8e92-32e26d404f37 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561697] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 50717fa4-956c-4e59-ab07-e6e8ec6f16bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561810] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance d2f84704-7d96-4ed6-835f-7176ff10148a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.561921] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b2aa6712-3284-4d97-ac98-ff8789fa8bca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.562046] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Migration 20ca0923-85a1-4eb9-89df-078708479759 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1111.562161] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 83369251-b00f-4595-bf98-28a3ec84f037 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.562318] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 55a1c5d3-7e00-40f6-9b06-f206bf2bf895 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1111.727479] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795303, 'name': CreateVM_Task, 'duration_secs': 0.358934} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.727659] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1111.728374] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.728548] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.728916] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1111.729257] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdbef7eb-4e80-4982-8516-26f215e8ca97 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.733631] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1111.733631] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52bbbdf5-68af-f024-2d5f-a68a53bfbb06" [ 1111.733631] env[61995]: _type = "Task" [ 1111.733631] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.741383] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52bbbdf5-68af-f024-2d5f-a68a53bfbb06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.824432] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b13d834-c35d-4f04-9365-e528ebfe8ebc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.833441] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e691189-54f4-4b57-b6ab-c94e92ffafaa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.897870] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.942527] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.942986] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.943173] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.944039] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab347b5-73ef-44d9-9967-045ddc44007e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.963944] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1111.964213] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1111.964376] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1111.964559] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1111.964710] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1111.964862] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1111.965085] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1111.965255] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1111.965432] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1111.965599] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1111.965771] env[61995]: DEBUG nova.virt.hardware [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1111.972804] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Reconfiguring VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1111.975561] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f43ff43-2c99-4621-8a14-2b391892d57f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.987986] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.988257] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.995591] env[61995]: DEBUG oslo_vmware.api [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1111.995591] env[61995]: value = "task-795304" [ 1111.995591] env[61995]: _type = "Task" [ 1111.995591] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.004299] env[61995]: DEBUG oslo_vmware.api [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795304, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.056710] env[61995]: DEBUG nova.network.neutron [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Updated VIF entry in instance network info cache for port e14467f7-216c-4d82-b986-9e488c9b78b6. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1112.057178] env[61995]: DEBUG nova.network.neutron [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Updating instance_info_cache with network_info: [{"id": "e14467f7-216c-4d82-b986-9e488c9b78b6", "address": "fa:16:3e:47:40:53", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape14467f7-21", "ovs_interfaceid": "e14467f7-216c-4d82-b986-9e488c9b78b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.065602] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 966783cb-b826-4983-9b77-32ab4ba56728 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1112.065843] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1112.065988] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1112.235287] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22375740-518e-47f9-b181-80f70663cfa9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.248722] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b11589-ae97-482a-9543-0b8f056925bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.251984] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.252260] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Processing image 258f374f-e693-46c8-87e6-9565d80e03ff {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1112.252502] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff/258f374f-e693-46c8-87e6-9565d80e03ff.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.252655] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff/258f374f-e693-46c8-87e6-9565d80e03ff.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.252843] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1112.254614] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b4831c7-6f0e-46cb-b5dd-e232d12978d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.280912] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb12ad8c-4cf6-473f-8536-34b501ae1ba7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.283540] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1112.283731] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1112.284460] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b855f68-511d-497e-b1a4-5721b5068641 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.289996] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1112.289996] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c7d61e-a27a-d292-c890-eab23dfc3083" [ 1112.289996] env[61995]: _type = "Task" [ 1112.289996] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.295336] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25ed49b-7973-41e8-8e04-579d0a94cf40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.303412] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c7d61e-a27a-d292-c890-eab23dfc3083, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.310829] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.491359] env[61995]: INFO nova.compute.manager [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Detaching volume ae89e4f5-9640-4273-a471-d85d8a021c3b [ 1112.508234] env[61995]: DEBUG oslo_vmware.api [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795304, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.524171] env[61995]: INFO nova.virt.block_device [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Attempting to driver detach volume ae89e4f5-9640-4273-a471-d85d8a021c3b from mountpoint /dev/sdb [ 1112.524440] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1112.524634] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185421', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'name': 'volume-ae89e4f5-9640-4273-a471-d85d8a021c3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e2014437-fbcd-454a-893f-9ad7c7d461d0', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'serial': 'ae89e4f5-9640-4273-a471-d85d8a021c3b'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1112.525503] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e76d5f3-09c1-4c53-b6f0-c8d78c380e4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.546669] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cdad70-cd6d-403a-8f01-0ab3d876e0a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.554726] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2090bf-f733-42e5-a3b9-75e19b2d1445 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.576067] env[61995]: DEBUG oslo_concurrency.lockutils [req-00992026-7367-4fc9-a137-5747a0de82ac req-85dbd6e8-6ff8-48ec-b63e-0b4522557878 service nova] Releasing lock "refresh_cache-55a1c5d3-7e00-40f6-9b06-f206bf2bf895" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.577458] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddfe2f4-2890-46ae-b3c3-57e0baf06872 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.592952] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] The volume has not been displaced from its original location: [datastore2] volume-ae89e4f5-9640-4273-a471-d85d8a021c3b/volume-ae89e4f5-9640-4273-a471-d85d8a021c3b.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1112.598147] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Reconfiguring VM instance instance-00000061 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1112.598470] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76fed293-5669-4258-98e6-2288bf8956b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.616805] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1112.616805] env[61995]: value = "task-795305" [ 1112.616805] env[61995]: _type = "Task" [ 1112.616805] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.624798] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795305, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.805107] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Preparing fetch location {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1112.805107] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Fetch image to [datastore2] OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f/OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f.vmdk {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1112.805282] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Downloading stream optimized image 258f374f-e693-46c8-87e6-9565d80e03ff to [datastore2] OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f/OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f.vmdk on the data store datastore2 as vApp {{(pid=61995) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1112.805434] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Downloading image file data 258f374f-e693-46c8-87e6-9565d80e03ff to the ESX as VM named 'OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f' {{(pid=61995) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1112.816580] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.881360] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1112.881360] env[61995]: value = "resgroup-9" [ 1112.881360] env[61995]: _type = "ResourcePool" [ 1112.881360] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1112.881975] env[61995]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-250ae184-fe83-4b9a-b6fc-c114d930a039 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.903205] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease: (returnval){ [ 1112.903205] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a2b4ea-be48-327a-5520-169f954a691c" [ 1112.903205] env[61995]: _type = "HttpNfcLease" [ 1112.903205] env[61995]: } obtained for vApp import into resource pool (val){ [ 1112.903205] env[61995]: value = "resgroup-9" [ 1112.903205] env[61995]: _type = "ResourcePool" [ 1112.903205] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1112.903465] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the lease: (returnval){ [ 1112.903465] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a2b4ea-be48-327a-5520-169f954a691c" [ 1112.903465] env[61995]: _type = "HttpNfcLease" [ 1112.903465] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1112.911561] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1112.911561] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a2b4ea-be48-327a-5520-169f954a691c" [ 1112.911561] env[61995]: _type = "HttpNfcLease" [ 1112.911561] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1112.933353] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1cf038-9b45-433f-928e-05c604dd6cb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.957237] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0960cae-0e2c-4054-851f-bca4df60d1a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.964350] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance '83369251-b00f-4595-bf98-28a3ec84f037' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1113.008226] env[61995]: DEBUG oslo_vmware.api [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795304, 'name': ReconfigVM_Task, 'duration_secs': 0.6569} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.008803] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.009035] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Reconfigured VM to attach interface {{(pid=61995) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1113.126225] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795305, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.321690] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1113.322023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.796s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.322309] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.424s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.323867] env[61995]: INFO nova.compute.claims [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1113.411496] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1113.411496] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a2b4ea-be48-327a-5520-169f954a691c" [ 1113.411496] env[61995]: _type = "HttpNfcLease" [ 1113.411496] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1113.411790] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1113.411790] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a2b4ea-be48-327a-5520-169f954a691c" [ 1113.411790] env[61995]: _type = "HttpNfcLease" [ 1113.411790] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1113.412497] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5845c3f5-2eff-4c5e-ac03-e29523e24e76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.419701] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529d1f00-b7e3-2c05-8a5c-7ef029fd4588/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1113.419895] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529d1f00-b7e3-2c05-8a5c-7ef029fd4588/disk-0.vmdk. {{(pid=61995) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1113.477632] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1113.477982] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29ec0b60-5076-4537-a1a0-e6fa3df7bc34 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.485113] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d03ae3ae-f01c-48ee-ab15-a879347fa823 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.489548] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1113.489548] env[61995]: value = "task-795307" [ 1113.489548] env[61995]: _type = "Task" [ 1113.489548] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.500112] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.513784] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d5bc65e8-045e-40a9-b55f-3359129aefc9 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-50717fa4-956c-4e59-ab07-e6e8ec6f16bf-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.592s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.627846] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795305, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.000484] env[61995]: DEBUG oslo_vmware.api [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795307, 'name': PowerOnVM_Task, 'duration_secs': 0.458067} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.002168] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1114.002359] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-11635272-5a83-4e4d-b1e6-b92fca5ea7da tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance '83369251-b00f-4595-bf98-28a3ec84f037' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1114.133058] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795305, 'name': ReconfigVM_Task, 'duration_secs': 1.342256} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.134407] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Reconfigured VM instance instance-00000061 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1114.139367] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df8e9cd6-f58c-46ed-9524-a3a9618d0ce3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.159933] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1114.159933] env[61995]: value = "task-795308" [ 1114.159933] env[61995]: _type = "Task" [ 1114.159933] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.171091] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795308, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.206236] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Completed reading data from the image iterator. {{(pid=61995) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1114.206573] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529d1f00-b7e3-2c05-8a5c-7ef029fd4588/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1114.207467] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439d51b9-a480-443d-8d9a-a8b31ffb2663 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.213656] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529d1f00-b7e3-2c05-8a5c-7ef029fd4588/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1114.213830] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529d1f00-b7e3-2c05-8a5c-7ef029fd4588/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1114.214066] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9843f7df-f418-46ee-94a2-b30a2eab3428 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.426013] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529d1f00-b7e3-2c05-8a5c-7ef029fd4588/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1114.426728] env[61995]: INFO nova.virt.vmwareapi.images [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Downloaded image file data 258f374f-e693-46c8-87e6-9565d80e03ff [ 1114.427782] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e039dd-d3f3-491a-90b7-306149080daa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.449651] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24a38321-7c1e-4216-be3b-6af9b6d15d9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.487076] env[61995]: INFO nova.virt.vmwareapi.images [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] The imported VM was unregistered [ 1114.489560] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Caching image {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1114.490607] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating directory with path [datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.490607] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d046ee6e-de2f-48fe-b68c-6ad926eefb8e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.504766] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created directory with path [datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.504962] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f/OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f.vmdk to [datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff/258f374f-e693-46c8-87e6-9565d80e03ff.vmdk. {{(pid=61995) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1114.505226] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-546ee140-d276-4387-9e32-bfd80663429d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.512555] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7869aa0c-9410-4a47-8376-414035ce4d1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.520383] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e723144-4ca2-4e77-8849-5cd2d60b12aa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.526151] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1114.526151] env[61995]: value = "task-795310" [ 1114.526151] env[61995]: _type = "Task" [ 1114.526151] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.558899] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca21f8a3-eb20-4e13-993a-76494bbe9b29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.561555] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795310, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.568138] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeefe342-029e-4d41-9899-1447dc443370 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.582918] env[61995]: DEBUG nova.compute.provider_tree [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.672315] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795308, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.037953] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795310, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.086329] env[61995]: DEBUG nova.scheduler.client.report [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.159055] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "interface-50717fa4-956c-4e59-ab07-e6e8ec6f16bf-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.159412] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-50717fa4-956c-4e59-ab07-e6e8ec6f16bf-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.176584] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795308, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.190801] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.191391] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.510818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.511187] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.540842] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795310, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.591476] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.592035] env[61995]: DEBUG nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1115.663226] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.663528] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.664458] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cafe2f13-eb4a-4daa-b102-c2c4920011c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.690235] env[61995]: DEBUG oslo_vmware.api [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795308, 'name': ReconfigVM_Task, 'duration_secs': 1.142649} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.691107] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185421', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'name': 'volume-ae89e4f5-9640-4273-a471-d85d8a021c3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e2014437-fbcd-454a-893f-9ad7c7d461d0', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae89e4f5-9640-4273-a471-d85d8a021c3b', 'serial': 'ae89e4f5-9640-4273-a471-d85d8a021c3b'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1115.693967] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f86752a-0f0e-4304-a11c-e976231c280a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.697598] env[61995]: DEBUG nova.compute.utils [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1115.725783] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Reconfiguring VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1115.727094] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59f95912-b71c-4db1-8906-16f3c66c483a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.746218] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1115.746218] env[61995]: value = "task-795311" [ 1115.746218] env[61995]: _type = "Task" [ 1115.746218] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.758652] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.888403] env[61995]: DEBUG nova.network.neutron [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Port c97b92c2-09c6-42ae-bac7-ad1583a82097 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1115.888705] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.888860] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.889040] env[61995]: DEBUG nova.network.neutron [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.014084] env[61995]: INFO nova.compute.manager [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Detaching volume 77915330-bd03-4a88-892c-dc7171c32e4a [ 1116.041451] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795310, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.055673] env[61995]: INFO nova.virt.block_device [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Attempting to driver detach volume 77915330-bd03-4a88-892c-dc7171c32e4a from mountpoint /dev/sdb [ 1116.055924] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1116.056101] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185408', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'name': 'volume-77915330-bd03-4a88-892c-dc7171c32e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66c56f7b-1fc7-4aed-8afc-350817e1ca48', 'attached_at': '', 'detached_at': '', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'serial': '77915330-bd03-4a88-892c-dc7171c32e4a'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1116.057037] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fadaa5a-f67c-4a00-8d37-45d593f94c62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.081386] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91534183-fdf5-43c2-bdce-87457ab28632 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.090950] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efc487c-c5b0-4dba-9ad3-3c66ceb2c539 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.097142] env[61995]: DEBUG nova.compute.utils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1116.116013] env[61995]: DEBUG nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1116.116240] env[61995]: DEBUG nova.network.neutron [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1116.119777] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8dfdb1-ccdd-4287-a218-3a4d7d713fa3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.139217] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] The volume has not been displaced from its original location: [datastore2] volume-77915330-bd03-4a88-892c-dc7171c32e4a/volume-77915330-bd03-4a88-892c-dc7171c32e4a.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1116.145796] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1116.146249] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9cc76d7-3556-4406-8522-1db379fc9d8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.168131] env[61995]: DEBUG oslo_vmware.api [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1116.168131] env[61995]: value = "task-795312" [ 1116.168131] env[61995]: _type = "Task" [ 1116.168131] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.172875] env[61995]: DEBUG nova.policy [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fdbcc3a4bbbb4a9db75e86dddf1dbc83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c776bce00156472db9b5f3d9e4205108', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1116.181391] env[61995]: DEBUG oslo_vmware.api [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795312, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.203399] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.240708] env[61995]: DEBUG nova.objects.instance [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'flavor' on Instance uuid e2014437-fbcd-454a-893f-9ad7c7d461d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.260497] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.475921] env[61995]: DEBUG nova.network.neutron [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Successfully created port: 011abc87-f621-435e-a18c-64d20e32d5f6 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1116.541138] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795310, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.600739] env[61995]: DEBUG nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1116.651011] env[61995]: DEBUG nova.network.neutron [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [{"id": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "address": "fa:16:3e:ae:d5:b8", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc97b92c2-09", "ovs_interfaceid": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.681302] env[61995]: DEBUG oslo_vmware.api [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795312, 'name': ReconfigVM_Task, 'duration_secs': 0.481029} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.681640] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1116.687062] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7341d06d-24e8-4955-88dd-20700c0d7ecd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.707954] env[61995]: DEBUG oslo_vmware.api [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1116.707954] env[61995]: value = "task-795313" [ 1116.707954] env[61995]: _type = "Task" [ 1116.707954] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.720767] env[61995]: DEBUG oslo_vmware.api [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795313, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.759681] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.038671] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795310, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.343122} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.038977] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f/OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f.vmdk to [datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff/258f374f-e693-46c8-87e6-9565d80e03ff.vmdk. [ 1117.039215] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Cleaning up location [datastore2] OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1117.039399] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_4c92368f-e969-4f1a-880a-cf00071a258f {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1117.039671] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c0fdd203-147e-43d1-8772-ebc7cf11cbcf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.046731] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1117.046731] env[61995]: value = "task-795314" [ 1117.046731] env[61995]: _type = "Task" [ 1117.046731] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.054186] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795314, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.157021] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.219292] env[61995]: DEBUG oslo_vmware.api [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795313, 'name': ReconfigVM_Task, 'duration_secs': 0.180791} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.219600] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185408', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'name': 'volume-77915330-bd03-4a88-892c-dc7171c32e4a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66c56f7b-1fc7-4aed-8afc-350817e1ca48', 'attached_at': '', 'detached_at': '', 'volume_id': '77915330-bd03-4a88-892c-dc7171c32e4a', 'serial': '77915330-bd03-4a88-892c-dc7171c32e4a'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1117.251782] env[61995]: DEBUG oslo_concurrency.lockutils [None req-d175611c-5ab3-4301-a250-50f1ccb5219c tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 5.263s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.258454] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.287621] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.287872] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.288140] env[61995]: INFO nova.compute.manager [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Attaching volume b7931f01-14ff-4e85-8499-8a899a9331a5 to /dev/sdb [ 1117.320549] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3b7956-b3c6-426e-b92b-6c724ee310ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.328085] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e094ec9-ef60-4ac7-9aa5-67610e98dff9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.341238] env[61995]: DEBUG nova.virt.block_device [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updating existing volume attachment record: e5666d7d-041d-408c-9ac5-764cfbc36df1 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1117.412429] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.412719] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.412942] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "e2014437-fbcd-454a-893f-9ad7c7d461d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.413154] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.413338] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.415683] env[61995]: INFO nova.compute.manager [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Terminating instance [ 1117.417593] env[61995]: DEBUG nova.compute.manager [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1117.417792] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1117.418827] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bfa2f1-e8a3-44f6-97f3-d207588c52e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.427529] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1117.427750] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0087f20c-26fc-43d8-9465-8cffd3672de7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.433675] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1117.433675] env[61995]: value = "task-795315" [ 1117.433675] env[61995]: _type = "Task" [ 1117.433675] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.441218] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795315, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.557130] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795314, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037162} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.557130] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.557130] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff/258f374f-e693-46c8-87e6-9565d80e03ff.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.557130] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff/258f374f-e693-46c8-87e6-9565d80e03ff.vmdk to [datastore2] 55a1c5d3-7e00-40f6-9b06-f206bf2bf895/55a1c5d3-7e00-40f6-9b06-f206bf2bf895.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1117.557130] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d9d7a845-8001-4cf6-b6aa-9bc35eff0b2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.565675] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1117.565675] env[61995]: value = "task-795317" [ 1117.565675] env[61995]: _type = "Task" [ 1117.565675] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.574203] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.614844] env[61995]: DEBUG nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1117.642495] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.642822] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.642988] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.643231] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.643418] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.643604] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.643903] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.644149] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.644383] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.644599] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.644860] env[61995]: DEBUG nova.virt.hardware [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.645709] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d43082c-b0b3-4df9-864a-b91de6b53436 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.654376] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa2d131-5809-4864-b477-c87eb9a3f445 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.660664] env[61995]: DEBUG nova.compute.manager [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61995) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1117.758666] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.762709] env[61995]: DEBUG nova.objects.instance [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lazy-loading 'flavor' on Instance uuid 66c56f7b-1fc7-4aed-8afc-350817e1ca48 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.946151] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795315, 'name': PowerOffVM_Task, 'duration_secs': 0.198605} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.946449] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1117.946625] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1117.946887] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-529cde1c-0bea-4eb3-a3a6-1f52a89338fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.064590] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1118.064920] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1118.065213] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleting the datastore file [datastore2] e2014437-fbcd-454a-893f-9ad7c7d461d0 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.065573] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4ba0c94-ba55-427f-9e9f-57eff43dba06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.079414] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795317, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.080975] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for the task: (returnval){ [ 1118.080975] env[61995]: value = "task-795319" [ 1118.080975] env[61995]: _type = "Task" [ 1118.080975] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.093729] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.192642] env[61995]: DEBUG nova.compute.manager [req-e492ad85-b6a1-47f4-9f9a-4f12d8df0542 req-7d91cb9a-3f99-458a-95ee-a0ada3279d64 service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Received event network-vif-plugged-011abc87-f621-435e-a18c-64d20e32d5f6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1118.193049] env[61995]: DEBUG oslo_concurrency.lockutils [req-e492ad85-b6a1-47f4-9f9a-4f12d8df0542 req-7d91cb9a-3f99-458a-95ee-a0ada3279d64 service nova] Acquiring lock "966783cb-b826-4983-9b77-32ab4ba56728-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.193421] env[61995]: DEBUG oslo_concurrency.lockutils [req-e492ad85-b6a1-47f4-9f9a-4f12d8df0542 req-7d91cb9a-3f99-458a-95ee-a0ada3279d64 service nova] Lock "966783cb-b826-4983-9b77-32ab4ba56728-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.193421] env[61995]: DEBUG oslo_concurrency.lockutils [req-e492ad85-b6a1-47f4-9f9a-4f12d8df0542 req-7d91cb9a-3f99-458a-95ee-a0ada3279d64 service nova] Lock "966783cb-b826-4983-9b77-32ab4ba56728-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.193421] env[61995]: DEBUG nova.compute.manager [req-e492ad85-b6a1-47f4-9f9a-4f12d8df0542 req-7d91cb9a-3f99-458a-95ee-a0ada3279d64 service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] No waiting events found dispatching network-vif-plugged-011abc87-f621-435e-a18c-64d20e32d5f6 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1118.193578] env[61995]: WARNING nova.compute.manager [req-e492ad85-b6a1-47f4-9f9a-4f12d8df0542 req-7d91cb9a-3f99-458a-95ee-a0ada3279d64 service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Received unexpected event network-vif-plugged-011abc87-f621-435e-a18c-64d20e32d5f6 for instance with vm_state building and task_state spawning. [ 1118.242420] env[61995]: DEBUG nova.network.neutron [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Successfully updated port: 011abc87-f621-435e-a18c-64d20e32d5f6 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1118.270455] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.581810] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795317, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.592607] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.745118] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "refresh_cache-966783cb-b826-4983-9b77-32ab4ba56728" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.745644] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquired lock "refresh_cache-966783cb-b826-4983-9b77-32ab4ba56728" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.745965] env[61995]: DEBUG nova.network.neutron [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1118.763875] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.773141] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3536c528-7f68-4b86-b088-6273c639b7be tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.262s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.870346] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.870796] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.080018] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795317, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.092201] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.265894] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.282356] env[61995]: DEBUG nova.network.neutron [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1119.373723] env[61995]: DEBUG nova.objects.instance [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'migration_context' on Instance uuid 83369251-b00f-4595-bf98-28a3ec84f037 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.476290] env[61995]: DEBUG nova.network.neutron [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Updating instance_info_cache with network_info: [{"id": "011abc87-f621-435e-a18c-64d20e32d5f6", "address": "fa:16:3e:08:c3:92", "network": {"id": "e991e74f-736e-4a4d-9fd3-d0c82d14f980", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-81578953-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c776bce00156472db9b5f3d9e4205108", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap011abc87-f6", "ovs_interfaceid": "011abc87-f621-435e-a18c-64d20e32d5f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.581026] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795317, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.593270] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.762302] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.958660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.960033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.960033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.960033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.960033] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.962031] env[61995]: INFO nova.compute.manager [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Terminating instance [ 1119.963761] env[61995]: DEBUG nova.compute.manager [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1119.964251] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1119.964848] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e134f9f1-09c4-4188-8e85-936e3b336917 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.972566] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1119.972807] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87013e52-3578-4492-9ef4-91c73f1f68e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.978868] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Releasing lock "refresh_cache-966783cb-b826-4983-9b77-32ab4ba56728" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.979187] env[61995]: DEBUG nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Instance network_info: |[{"id": "011abc87-f621-435e-a18c-64d20e32d5f6", "address": "fa:16:3e:08:c3:92", "network": {"id": "e991e74f-736e-4a4d-9fd3-d0c82d14f980", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-81578953-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c776bce00156472db9b5f3d9e4205108", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap011abc87-f6", "ovs_interfaceid": "011abc87-f621-435e-a18c-64d20e32d5f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1119.979540] env[61995]: DEBUG oslo_vmware.api [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1119.979540] env[61995]: value = "task-795321" [ 1119.979540] env[61995]: _type = "Task" [ 1119.979540] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.981976] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:c3:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '011abc87-f621-435e-a18c-64d20e32d5f6', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1119.989931] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Creating folder: Project (c776bce00156472db9b5f3d9e4205108). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1119.990556] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52af620f-356d-42d2-9232-c97d9d157b8c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.002432] env[61995]: DEBUG oslo_vmware.api [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.006707] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Created folder: Project (c776bce00156472db9b5f3d9e4205108) in parent group-v185203. [ 1120.006942] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Creating folder: Instances. Parent ref: group-v185425. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1120.007211] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07408d7c-d28d-4c50-8fd7-8130bea4d5f5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.017877] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Created folder: Instances in parent group-v185425. [ 1120.018152] env[61995]: DEBUG oslo.service.loopingcall [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1120.018361] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1120.018563] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b3700ee-83a0-4dbf-9c83-750283c2a044 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.041421] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1120.041421] env[61995]: value = "task-795324" [ 1120.041421] env[61995]: _type = "Task" [ 1120.041421] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.049693] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795324, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.079489] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795317, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.395463} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.082111] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/258f374f-e693-46c8-87e6-9565d80e03ff/258f374f-e693-46c8-87e6-9565d80e03ff.vmdk to [datastore2] 55a1c5d3-7e00-40f6-9b06-f206bf2bf895/55a1c5d3-7e00-40f6-9b06-f206bf2bf895.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1120.083139] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b484053-2159-4c4d-a9b3-2e668edde822 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.112365] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 55a1c5d3-7e00-40f6-9b06-f206bf2bf895/55a1c5d3-7e00-40f6-9b06-f206bf2bf895.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.112930] env[61995]: DEBUG oslo_vmware.api [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Task: {'id': task-795319, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.627694} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.114244] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53bc5877-ef2a-4d7d-bafd-249dfa8783f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.129493] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1120.129761] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1120.129991] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1120.130202] env[61995]: INFO nova.compute.manager [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Took 2.71 seconds to destroy the instance on the hypervisor. [ 1120.130489] env[61995]: DEBUG oslo.service.loopingcall [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1120.131236] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41f242d-cbe8-49b9-a104-488b41b5fe98 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.133945] env[61995]: DEBUG nova.compute.manager [-] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1120.134060] env[61995]: DEBUG nova.network.neutron [-] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1120.141899] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ad5d43-5828-40e7-87df-0131bc2cd61d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.145079] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1120.145079] env[61995]: value = "task-795325" [ 1120.145079] env[61995]: _type = "Task" [ 1120.145079] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.175731] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d790e7-5202-4f43-9d83-d22f834e4afa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.178751] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795325, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.186939] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfde89d-0d66-477d-9ea4-9fe9da1d111d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.202668] env[61995]: DEBUG nova.compute.provider_tree [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.236204] env[61995]: DEBUG nova.compute.manager [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Received event network-changed-011abc87-f621-435e-a18c-64d20e32d5f6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1120.236296] env[61995]: DEBUG nova.compute.manager [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Refreshing instance network info cache due to event network-changed-011abc87-f621-435e-a18c-64d20e32d5f6. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1120.237603] env[61995]: DEBUG oslo_concurrency.lockutils [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] Acquiring lock "refresh_cache-966783cb-b826-4983-9b77-32ab4ba56728" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.237603] env[61995]: DEBUG oslo_concurrency.lockutils [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] Acquired lock "refresh_cache-966783cb-b826-4983-9b77-32ab4ba56728" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.237603] env[61995]: DEBUG nova.network.neutron [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Refreshing network info cache for port 011abc87-f621-435e-a18c-64d20e32d5f6 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1120.260564] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.499810] env[61995]: DEBUG oslo_vmware.api [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795321, 'name': PowerOffVM_Task, 'duration_secs': 0.375779} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.500532] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1120.500532] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1120.500717] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9361456f-2f7d-42f3-aaf0-b4e922bb200f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.551082] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795324, 'name': CreateVM_Task, 'duration_secs': 0.3932} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.551260] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1120.551950] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.552145] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.552512] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1120.552777] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81853469-feeb-4ca5-8edb-7b85aed39f1b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.557129] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1120.557129] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f2d209-34a5-295b-2806-fa268ef3bf8c" [ 1120.557129] env[61995]: _type = "Task" [ 1120.557129] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.564713] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f2d209-34a5-295b-2806-fa268ef3bf8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.593176] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1120.593357] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1120.593594] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleting the datastore file [datastore2] 66c56f7b-1fc7-4aed-8afc-350817e1ca48 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1120.594593] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e44f3afb-3f6f-49a9-a925-155b73b136ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.600525] env[61995]: DEBUG oslo_vmware.api [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1120.600525] env[61995]: value = "task-795327" [ 1120.600525] env[61995]: _type = "Task" [ 1120.600525] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.608612] env[61995]: DEBUG oslo_vmware.api [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795327, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.659041] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795325, 'name': ReconfigVM_Task, 'duration_secs': 0.382624} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.659364] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 55a1c5d3-7e00-40f6-9b06-f206bf2bf895/55a1c5d3-7e00-40f6-9b06-f206bf2bf895.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.660192] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bb087c5-736d-4adb-87b2-acaf8db7b207 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.667152] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1120.667152] env[61995]: value = "task-795328" [ 1120.667152] env[61995]: _type = "Task" [ 1120.667152] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.676258] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795328, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.706605] env[61995]: DEBUG nova.scheduler.client.report [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.762920] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.934918] env[61995]: DEBUG nova.network.neutron [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Updated VIF entry in instance network info cache for port 011abc87-f621-435e-a18c-64d20e32d5f6. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1120.935422] env[61995]: DEBUG nova.network.neutron [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Updating instance_info_cache with network_info: [{"id": "011abc87-f621-435e-a18c-64d20e32d5f6", "address": "fa:16:3e:08:c3:92", "network": {"id": "e991e74f-736e-4a4d-9fd3-d0c82d14f980", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-81578953-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c776bce00156472db9b5f3d9e4205108", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap011abc87-f6", "ovs_interfaceid": "011abc87-f621-435e-a18c-64d20e32d5f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.067857] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f2d209-34a5-295b-2806-fa268ef3bf8c, 'name': SearchDatastore_Task, 'duration_secs': 0.0099} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.068224] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.068482] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1121.068729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.068882] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.069075] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1121.069385] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0af1d4b6-2c5d-44d8-95b6-9d722903ef07 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.077222] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1121.077415] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1121.078159] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43566f06-344c-49d0-b9e3-dad08f72b7a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.083365] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1121.083365] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]524b0b7f-0ada-b8a1-d312-a5643cca9023" [ 1121.083365] env[61995]: _type = "Task" [ 1121.083365] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.092191] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524b0b7f-0ada-b8a1-d312-a5643cca9023, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.110945] env[61995]: DEBUG oslo_vmware.api [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795327, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160201} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.111217] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1121.111417] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1121.111601] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1121.111838] env[61995]: INFO nova.compute.manager [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1121.112207] env[61995]: DEBUG oslo.service.loopingcall [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1121.112484] env[61995]: DEBUG nova.compute.manager [-] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1121.112639] env[61995]: DEBUG nova.network.neutron [-] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1121.154994] env[61995]: DEBUG nova.network.neutron [-] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.178327] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795328, 'name': Rename_Task, 'duration_secs': 0.158442} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.178665] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1121.178948] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d583cb5f-b4b4-4291-a5e9-a0d52594ba1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.185185] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1121.185185] env[61995]: value = "task-795329" [ 1121.185185] env[61995]: _type = "Task" [ 1121.185185] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.193643] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795329, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.266777] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.439122] env[61995]: DEBUG oslo_concurrency.lockutils [req-1bd623ac-d3ec-45ba-a0f8-560984702006 req-5c0c0ec8-46b7-45bb-b816-e9ac50655acc service nova] Releasing lock "refresh_cache-966783cb-b826-4983-9b77-32ab4ba56728" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.595283] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524b0b7f-0ada-b8a1-d312-a5643cca9023, 'name': SearchDatastore_Task, 'duration_secs': 0.008721} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.596279] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d69cd69e-aa57-43b8-b128-979b0fe6d85d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.603688] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1121.603688] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525fda0a-cb35-c168-96b9-ceecb1c77d44" [ 1121.603688] env[61995]: _type = "Task" [ 1121.603688] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.611878] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525fda0a-cb35-c168-96b9-ceecb1c77d44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.656872] env[61995]: INFO nova.compute.manager [-] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Took 1.52 seconds to deallocate network for instance. [ 1121.695126] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795329, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.718221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.847s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.763763] env[61995]: DEBUG oslo_vmware.api [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795311, 'name': ReconfigVM_Task, 'duration_secs': 5.793119} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.763974] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.764213] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Reconfigured VM to detach interface {{(pid=61995) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1121.885063] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1121.885334] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185424', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'name': 'volume-b7931f01-14ff-4e85-8499-8a899a9331a5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'serial': 'b7931f01-14ff-4e85-8499-8a899a9331a5'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1121.886351] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a18793c-5226-4960-9a04-a8185d4634a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.903534] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e4a1e4-40fe-4ce0-88f3-8485b46c5f34 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.928916] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] volume-b7931f01-14ff-4e85-8499-8a899a9331a5/volume-b7931f01-14ff-4e85-8499-8a899a9331a5.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1121.929273] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92078b64-5fe1-4aff-a826-b95a5ae70c9a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.949549] env[61995]: DEBUG oslo_vmware.api [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1121.949549] env[61995]: value = "task-795330" [ 1121.949549] env[61995]: _type = "Task" [ 1121.949549] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.961384] env[61995]: DEBUG oslo_vmware.api [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795330, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.114435] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525fda0a-cb35-c168-96b9-ceecb1c77d44, 'name': SearchDatastore_Task, 'duration_secs': 0.010502} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.114748] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.115064] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 966783cb-b826-4983-9b77-32ab4ba56728/966783cb-b826-4983-9b77-32ab4ba56728.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1122.115423] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1f4aab2-8f7d-4629-b438-3c4c2507bcb7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.121928] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1122.121928] env[61995]: value = "task-795331" [ 1122.121928] env[61995]: _type = "Task" [ 1122.121928] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.129985] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795331, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.140947] env[61995]: DEBUG nova.network.neutron [-] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.163522] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.163832] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.164088] env[61995]: DEBUG nova.objects.instance [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lazy-loading 'resources' on Instance uuid e2014437-fbcd-454a-893f-9ad7c7d461d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.196286] env[61995]: DEBUG oslo_vmware.api [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795329, 'name': PowerOnVM_Task, 'duration_secs': 0.792549} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.196659] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1122.196910] env[61995]: INFO nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Took 13.30 seconds to spawn the instance on the hypervisor. [ 1122.197157] env[61995]: DEBUG nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1122.198034] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18df58c8-ab41-4f83-9358-021920ab3e65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.276052] env[61995]: DEBUG nova.compute.manager [req-b9525e7e-93fc-4bcb-8e0d-2fdcbf3471b7 req-a56fb1ad-39df-4ca7-8aaf-8232f4087ab3 service nova] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Received event network-vif-deleted-534906ba-0f5d-4b75-91c1-ac34f5afd2ff {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1122.276932] env[61995]: DEBUG nova.compute.manager [req-b9525e7e-93fc-4bcb-8e0d-2fdcbf3471b7 req-a56fb1ad-39df-4ca7-8aaf-8232f4087ab3 service nova] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Received event network-vif-deleted-ff48bd88-ebef-44df-92d3-1a4b9c6ff71d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1122.461059] env[61995]: DEBUG oslo_vmware.api [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795330, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.631753] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795331, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4883} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.632026] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 966783cb-b826-4983-9b77-32ab4ba56728/966783cb-b826-4983-9b77-32ab4ba56728.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1122.632252] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1122.632507] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bafc936-5f1f-4bcc-bf38-19ad5184fb2b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.639662] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1122.639662] env[61995]: value = "task-795332" [ 1122.639662] env[61995]: _type = "Task" [ 1122.639662] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.643122] env[61995]: INFO nova.compute.manager [-] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Took 1.53 seconds to deallocate network for instance. [ 1122.647690] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795332, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.717251] env[61995]: INFO nova.compute.manager [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Took 18.10 seconds to build instance. [ 1122.833512] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41897b8b-5bbc-4d4d-9d1e-37a2979a1b73 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.841569] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62228c4b-ba9d-4cef-b9ce-ef42194eca6e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.870755] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4482f685-a806-4869-a923-5a27656d4d97 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.878052] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2476a8a8-af30-407c-865b-e79863d19389 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.890634] env[61995]: DEBUG nova.compute.provider_tree [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.960459] env[61995]: DEBUG oslo_vmware.api [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795330, 'name': ReconfigVM_Task, 'duration_secs': 0.548516} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.960793] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfigured VM instance instance-00000063 to attach disk [datastore2] volume-b7931f01-14ff-4e85-8499-8a899a9331a5/volume-b7931f01-14ff-4e85-8499-8a899a9331a5.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1122.965637] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3013995e-2716-4b82-86ff-eefc68c44ddf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.979739] env[61995]: DEBUG oslo_vmware.api [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1122.979739] env[61995]: value = "task-795333" [ 1122.979739] env[61995]: _type = "Task" [ 1122.979739] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.987182] env[61995]: DEBUG oslo_vmware.api [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.066379] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.066565] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquired lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.066741] env[61995]: DEBUG nova.network.neutron [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.152507] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.152842] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795332, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070661} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.153134] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1123.153977] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859b2e36-7b96-4494-847f-456284ae863f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.178017] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 966783cb-b826-4983-9b77-32ab4ba56728/966783cb-b826-4983-9b77-32ab4ba56728.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1123.178373] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca8bc459-21d7-44a2-9167-d2a9f9fc4e64 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.199824] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1123.199824] env[61995]: value = "task-795334" [ 1123.199824] env[61995]: _type = "Task" [ 1123.199824] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.208639] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.219443] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6d03c131-c922-4639-b57c-6b4af8b01192 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.615s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.255578] env[61995]: INFO nova.compute.manager [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Swapping old allocation on dict_keys(['5c086f4d-bc91-4e49-9831-bed8df133c15']) held by migration 20ca0923-85a1-4eb9-89df-078708479759 for instance [ 1123.278069] env[61995]: DEBUG nova.scheduler.client.report [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Overwriting current allocation {'allocations': {'5c086f4d-bc91-4e49-9831-bed8df133c15': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 140}}, 'project_id': 'cd23f8abd8f14ec392fbfb7fd5bc64f9', 'user_id': '4fa18e6b6fe742909fa15a846d3019e6', 'consumer_generation': 1} on consumer 83369251-b00f-4595-bf98-28a3ec84f037 {{(pid=61995) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1123.371770] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.371973] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.372177] env[61995]: DEBUG nova.network.neutron [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.393746] env[61995]: DEBUG nova.scheduler.client.report [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1123.494694] env[61995]: DEBUG oslo_vmware.api [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795333, 'name': ReconfigVM_Task, 'duration_secs': 0.150845} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.495015] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185424', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'name': 'volume-b7931f01-14ff-4e85-8499-8a899a9331a5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'serial': 'b7931f01-14ff-4e85-8499-8a899a9331a5'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1123.557867] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.557867] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.558666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.558666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.558666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.561689] env[61995]: INFO nova.compute.manager [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Terminating instance [ 1123.564300] env[61995]: DEBUG nova.compute.manager [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1123.564300] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1123.564942] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3630cc-4c1f-4b1b-bdf6-5384f31e21ab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.574253] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1123.574868] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-013c1141-aa8c-4944-9516-ecaf8032b8d0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.581884] env[61995]: DEBUG oslo_vmware.api [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1123.581884] env[61995]: value = "task-795335" [ 1123.581884] env[61995]: _type = "Task" [ 1123.581884] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.590634] env[61995]: DEBUG oslo_vmware.api [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795335, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.710544] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795334, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.819329] env[61995]: INFO nova.network.neutron [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Port 5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1123.819734] env[61995]: DEBUG nova.network.neutron [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [{"id": "3cd23361-0e98-4153-b3ea-06ce74356da6", "address": "fa:16:3e:ae:91:30", "network": {"id": "67eef3c3-44a6-48d3-9548-e77c76e98379", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1779718710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a660ec6d4d7e4e76827642cf247f53c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd23361-0e", "ovs_interfaceid": "3cd23361-0e98-4153-b3ea-06ce74356da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.899420] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.735s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.901779] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.749s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.902035] env[61995]: DEBUG nova.objects.instance [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lazy-loading 'resources' on Instance uuid 66c56f7b-1fc7-4aed-8afc-350817e1ca48 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.917907] env[61995]: INFO nova.scheduler.client.report [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Deleted allocations for instance e2014437-fbcd-454a-893f-9ad7c7d461d0 [ 1123.990295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.990562] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.990779] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.990972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.991166] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.993186] env[61995]: INFO nova.compute.manager [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Terminating instance [ 1123.995070] env[61995]: DEBUG nova.compute.manager [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1123.995268] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1123.996116] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b985bf-ae07-4557-9ecc-3528f87a1602 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.005722] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1124.005957] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0da7e468-82f2-4e38-bce0-150dea0b07a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.012244] env[61995]: DEBUG oslo_vmware.api [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1124.012244] env[61995]: value = "task-795336" [ 1124.012244] env[61995]: _type = "Task" [ 1124.012244] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.023989] env[61995]: DEBUG oslo_vmware.api [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795336, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.091618] env[61995]: DEBUG oslo_vmware.api [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795335, 'name': PowerOffVM_Task, 'duration_secs': 0.38722} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.091919] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1124.092137] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1124.092422] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e28b3cf-2e60-4466-b3ed-0921bd608ba0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.112628] env[61995]: DEBUG nova.network.neutron [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [{"id": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "address": "fa:16:3e:ae:d5:b8", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc97b92c2-09", "ovs_interfaceid": "c97b92c2-09c6-42ae-bac7-ad1583a82097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.168516] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1124.168855] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1124.169132] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleting the datastore file [datastore2] 50717fa4-956c-4e59-ab07-e6e8ec6f16bf {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1124.169460] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25e7c3f6-2e51-4cee-9c0c-b2fd4cbfe460 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.176668] env[61995]: DEBUG oslo_vmware.api [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1124.176668] env[61995]: value = "task-795338" [ 1124.176668] env[61995]: _type = "Task" [ 1124.176668] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.184725] env[61995]: DEBUG oslo_vmware.api [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.208493] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795334, 'name': ReconfigVM_Task, 'duration_secs': 0.791527} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.208782] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 966783cb-b826-4983-9b77-32ab4ba56728/966783cb-b826-4983-9b77-32ab4ba56728.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.209466] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddd7dda1-1ea0-44e1-8edb-216ca4ae6b1c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.215236] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1124.215236] env[61995]: value = "task-795339" [ 1124.215236] env[61995]: _type = "Task" [ 1124.215236] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.222945] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795339, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.322492] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Releasing lock "refresh_cache-50717fa4-956c-4e59-ab07-e6e8ec6f16bf" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.426043] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80a7d4df-40ec-41f6-92d8-8266f3a84a1e tempest-AttachVolumeNegativeTest-1177146387 tempest-AttachVolumeNegativeTest-1177146387-project-member] Lock "e2014437-fbcd-454a-893f-9ad7c7d461d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.013s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.523816] env[61995]: DEBUG oslo_vmware.api [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795336, 'name': PowerOffVM_Task, 'duration_secs': 0.147539} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.524118] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1124.524299] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1124.524850] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9e83c53-4f06-4202-82a0-57b183cc55b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.531721] env[61995]: DEBUG nova.objects.instance [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid a73742da-3d5c-4b71-8e92-32e26d404f37 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.539130] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf45f03-6274-415b-aaa1-93b78d19fadf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.546763] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83d4073-5c90-4294-bf5f-062b63997051 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.578270] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ebc70d-57c2-4b74-9b11-57995d36528a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.587504] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fb6403-5ea9-4568-8250-d3a248d69ff7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.591680] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1124.591891] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1124.592094] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleting the datastore file [datastore2] 55a1c5d3-7e00-40f6-9b06-f206bf2bf895 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1124.592332] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d0fd61dc-d8c7-4590-bfd3-11ff53796e70 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.604317] env[61995]: DEBUG nova.compute.provider_tree [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.606821] env[61995]: DEBUG oslo_vmware.api [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1124.606821] env[61995]: value = "task-795341" [ 1124.606821] env[61995]: _type = "Task" [ 1124.606821] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.617809] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-83369251-b00f-4595-bf98-28a3ec84f037" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.618462] env[61995]: DEBUG oslo_vmware.api [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795341, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.619058] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e637a1f8-be0b-4fb8-b083-016408f705f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.626294] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c11f1a8-80ab-476a-954b-08ca36779bc9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.686663] env[61995]: DEBUG oslo_vmware.api [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144398} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.686937] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1124.687148] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1124.687331] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1124.687513] env[61995]: INFO nova.compute.manager [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1124.688030] env[61995]: DEBUG oslo.service.loopingcall [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.688030] env[61995]: DEBUG nova.compute.manager [-] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1124.688229] env[61995]: DEBUG nova.network.neutron [-] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1124.725688] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795339, 'name': Rename_Task, 'duration_secs': 0.146174} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.727099] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1124.727099] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f245bc8c-70a1-4b78-880c-077ed7ad71db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.732757] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1124.732757] env[61995]: value = "task-795342" [ 1124.732757] env[61995]: _type = "Task" [ 1124.732757] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.743480] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795342, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.826373] env[61995]: DEBUG oslo_concurrency.lockutils [None req-5b5d01e1-960c-4a5b-9775-639a8c3d2d65 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "interface-50717fa4-956c-4e59-ab07-e6e8ec6f16bf-5b27da3c-fcc5-4a26-b2d3-10f0b98ecdb1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.667s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.036340] env[61995]: DEBUG oslo_concurrency.lockutils [None req-8b2b9924-211a-450b-87b3-b39d1859dc70 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.748s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.108743] env[61995]: DEBUG nova.scheduler.client.report [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.121968] env[61995]: DEBUG oslo_vmware.api [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795341, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128259} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.122400] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1125.122741] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1125.123080] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1125.123354] env[61995]: INFO nova.compute.manager [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1125.123753] env[61995]: DEBUG oslo.service.loopingcall [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.124071] env[61995]: DEBUG nova.compute.manager [-] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1125.124439] env[61995]: DEBUG nova.network.neutron [-] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1125.245258] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795342, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.512978] env[61995]: DEBUG nova.compute.manager [req-8ac5d063-9694-4f29-9d41-252609941857 req-e1bf5cb7-6879-4ba6-80ea-91fce7ff88e8 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Received event network-vif-deleted-e14467f7-216c-4d82-b986-9e488c9b78b6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.513262] env[61995]: INFO nova.compute.manager [req-8ac5d063-9694-4f29-9d41-252609941857 req-e1bf5cb7-6879-4ba6-80ea-91fce7ff88e8 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Neutron deleted interface e14467f7-216c-4d82-b986-9e488c9b78b6; detaching it from the instance and deleting it from the info cache [ 1125.513480] env[61995]: DEBUG nova.network.neutron [req-8ac5d063-9694-4f29-9d41-252609941857 req-e1bf5cb7-6879-4ba6-80ea-91fce7ff88e8 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.566741] env[61995]: DEBUG nova.compute.manager [req-1d4d39bb-eb53-4029-80ab-4a2fb183bb84 req-9b7dc22d-4e27-4e26-9c48-6a67c4d9de33 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Received event network-vif-deleted-3cd23361-0e98-4153-b3ea-06ce74356da6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.566967] env[61995]: INFO nova.compute.manager [req-1d4d39bb-eb53-4029-80ab-4a2fb183bb84 req-9b7dc22d-4e27-4e26-9c48-6a67c4d9de33 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Neutron deleted interface 3cd23361-0e98-4153-b3ea-06ce74356da6; detaching it from the instance and deleting it from the info cache [ 1125.567176] env[61995]: DEBUG nova.network.neutron [req-1d4d39bb-eb53-4029-80ab-4a2fb183bb84 req-9b7dc22d-4e27-4e26-9c48-6a67c4d9de33 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.617911] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.715s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.640371] env[61995]: INFO nova.scheduler.client.report [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleted allocations for instance 66c56f7b-1fc7-4aed-8afc-350817e1ca48 [ 1125.720155] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1125.720155] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9682a420-41aa-4b25-ad65-ba411654b559 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.727021] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1125.727021] env[61995]: value = "task-795344" [ 1125.727021] env[61995]: _type = "Task" [ 1125.727021] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.738067] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795344, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.748619] env[61995]: DEBUG oslo_vmware.api [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795342, 'name': PowerOnVM_Task, 'duration_secs': 0.783668} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.749092] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1125.749414] env[61995]: INFO nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Took 8.13 seconds to spawn the instance on the hypervisor. [ 1125.749703] env[61995]: DEBUG nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1125.750600] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eee8250-44dc-402c-83b4-7ca7e6f78ae8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.989579] env[61995]: DEBUG nova.network.neutron [-] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.993475] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.993646] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.015907] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b1d861f-2bf5-4f4b-bbf7-a29b0d925a6a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.025222] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b4a4c5-4c8f-4810-94cb-c79a7d74f177 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.045864] env[61995]: DEBUG nova.network.neutron [-] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.056782] env[61995]: DEBUG nova.compute.manager [req-8ac5d063-9694-4f29-9d41-252609941857 req-e1bf5cb7-6879-4ba6-80ea-91fce7ff88e8 service nova] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Detach interface failed, port_id=e14467f7-216c-4d82-b986-9e488c9b78b6, reason: Instance 55a1c5d3-7e00-40f6-9b06-f206bf2bf895 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1126.069434] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e124c776-0565-40c4-ae73-116d45020a96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.078873] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef2f251-67d6-42a1-a09e-c54febd2a02e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.106539] env[61995]: DEBUG nova.compute.manager [req-1d4d39bb-eb53-4029-80ab-4a2fb183bb84 req-9b7dc22d-4e27-4e26-9c48-6a67c4d9de33 service nova] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Detach interface failed, port_id=3cd23361-0e98-4153-b3ea-06ce74356da6, reason: Instance 50717fa4-956c-4e59-ab07-e6e8ec6f16bf could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1126.150501] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1f31ea04-98f7-4853-9a16-67eda13b53d1 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "66c56f7b-1fc7-4aed-8afc-350817e1ca48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.191s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.238163] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795344, 'name': PowerOffVM_Task, 'duration_secs': 0.22138} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.238543] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1126.239292] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1126.239511] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1126.239683] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1126.239885] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1126.240058] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1126.240220] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1126.240444] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1126.240612] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1126.240786] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1126.240956] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1126.241162] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.246724] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13941505-8d1c-4c16-a0dc-b576646cea55 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.263532] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1126.263532] env[61995]: value = "task-795345" [ 1126.263532] env[61995]: _type = "Task" [ 1126.263532] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.272558] env[61995]: INFO nova.compute.manager [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Took 14.39 seconds to build instance. [ 1126.277644] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795345, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.493443] env[61995]: INFO nova.compute.manager [-] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Took 1.37 seconds to deallocate network for instance. [ 1126.500735] env[61995]: DEBUG nova.compute.utils [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1126.558266] env[61995]: INFO nova.compute.manager [-] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Took 1.87 seconds to deallocate network for instance. [ 1126.773838] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795345, 'name': ReconfigVM_Task, 'duration_secs': 0.165425} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.774269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c5aeb94c-c16b-4334-8f7e-881b68e9111a tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "966783cb-b826-4983-9b77-32ab4ba56728" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.904s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.775082] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c682f269-dea2-4d50-a82d-4a4f64e74f60 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.797895] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1126.798261] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1126.798465] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1126.798667] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1126.798821] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1126.798973] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1126.799254] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1126.799428] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1126.799607] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1126.799782] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1126.799962] env[61995]: DEBUG nova.virt.hardware [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.800768] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9306e4e4-2406-4bbc-96d3-745066ef8710 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.805978] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1126.805978] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52eada93-36b7-5583-828c-b47da3c3a2b4" [ 1126.805978] env[61995]: _type = "Task" [ 1126.805978] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.813665] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52eada93-36b7-5583-828c-b47da3c3a2b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.005436] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.005436] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.005436] env[61995]: DEBUG nova.objects.instance [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lazy-loading 'resources' on Instance uuid 55a1c5d3-7e00-40f6-9b06-f206bf2bf895 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.006117] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.065538] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.315634] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52eada93-36b7-5583-828c-b47da3c3a2b4, 'name': SearchDatastore_Task, 'duration_secs': 0.007177} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.320974] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1127.321590] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae6afee6-6c0c-49df-8eb8-7053672d5dc0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.339623] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1127.339623] env[61995]: value = "task-795347" [ 1127.339623] env[61995]: _type = "Task" [ 1127.339623] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.349467] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795347, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.652548] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8956a48-0bee-4966-94c8-e4141c986583 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.659885] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5063fba-ecb2-447b-9a02-01152aef2470 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.692027] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104d84d8-19c2-40cc-9506-4da4182abc8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.696566] env[61995]: DEBUG nova.compute.manager [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1127.711270] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d18c3c4-5224-407a-8b41-958e7d22a635 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.726158] env[61995]: DEBUG nova.compute.provider_tree [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.849595] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795347, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.004280] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "966783cb-b826-4983-9b77-32ab4ba56728" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.004587] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "966783cb-b826-4983-9b77-32ab4ba56728" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.004801] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "966783cb-b826-4983-9b77-32ab4ba56728-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.004996] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "966783cb-b826-4983-9b77-32ab4ba56728-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.005192] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "966783cb-b826-4983-9b77-32ab4ba56728-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.007528] env[61995]: INFO nova.compute.manager [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Terminating instance [ 1128.009524] env[61995]: DEBUG nova.compute.manager [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1128.009729] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1128.010596] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e1a2a1-378f-4993-884f-49134818b5fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.018284] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1128.018489] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b8173b1-3f3d-436e-a4ed-9f444b70a523 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.025263] env[61995]: DEBUG oslo_vmware.api [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1128.025263] env[61995]: value = "task-795348" [ 1128.025263] env[61995]: _type = "Task" [ 1128.025263] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.032849] env[61995]: DEBUG oslo_vmware.api [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795348, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.060866] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.061199] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.061474] env[61995]: INFO nova.compute.manager [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Attaching volume 17f21f6a-4284-4d68-b7ee-6d82e0d56cdc to /dev/sdc [ 1128.093112] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29926126-913e-451d-b93d-08150d57ca29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.100837] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46957668-a300-4342-8eed-03d85d47d98b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.113934] env[61995]: DEBUG nova.virt.block_device [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updating existing volume attachment record: 40613508-c523-4462-b182-c97a582b1454 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1128.213485] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.228999] env[61995]: DEBUG nova.scheduler.client.report [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.349752] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795347, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.534572] env[61995]: DEBUG oslo_vmware.api [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795348, 'name': PowerOffVM_Task, 'duration_secs': 0.189184} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.534851] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1128.535036] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1128.535361] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b8e7463-42ad-4a40-97d7-008ca107a073 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.661261] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1128.661592] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1128.661839] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Deleting the datastore file [datastore2] 966783cb-b826-4983-9b77-32ab4ba56728 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1128.662178] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a6506b8-a71a-4040-b238-6f2c7caeadf3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.668020] env[61995]: DEBUG oslo_vmware.api [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for the task: (returnval){ [ 1128.668020] env[61995]: value = "task-795352" [ 1128.668020] env[61995]: _type = "Task" [ 1128.668020] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.677158] env[61995]: DEBUG oslo_vmware.api [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795352, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.735074] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.738414] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.672s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.738704] env[61995]: DEBUG nova.objects.instance [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'resources' on Instance uuid 50717fa4-956c-4e59-ab07-e6e8ec6f16bf {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.756941] env[61995]: INFO nova.scheduler.client.report [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted allocations for instance 55a1c5d3-7e00-40f6-9b06-f206bf2bf895 [ 1128.851060] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795347, 'name': ReconfigVM_Task, 'duration_secs': 1.27283} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.851423] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1128.852085] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b64b3ca1-133f-4a7a-9d39-06caab646b62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.877352] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.877658] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90b269ef-ef1a-4c8d-a71f-c221729eb470 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.894702] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1128.894702] env[61995]: value = "task-795353" [ 1128.894702] env[61995]: _type = "Task" [ 1128.894702] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.904774] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795353, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.178824] env[61995]: DEBUG oslo_vmware.api [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Task: {'id': task-795352, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128402} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.179181] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.179410] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1129.179605] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1129.179786] env[61995]: INFO nova.compute.manager [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1129.180049] env[61995]: DEBUG oslo.service.loopingcall [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.180292] env[61995]: DEBUG nova.compute.manager [-] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1129.180414] env[61995]: DEBUG nova.network.neutron [-] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1129.266994] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e0fc7ecc-b0cd-492b-b2fb-0bd1093dfa11 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "55a1c5d3-7e00-40f6-9b06-f206bf2bf895" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.276s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.383048] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fade0cfa-bc87-4b42-8960-16e45a71e1d1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.390445] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820db59d-d58c-44a5-a0c2-27cd749907ca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.426300] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2cc56d-dbf5-4bf5-bb4d-afc8f1ae4e59 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.428850] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795353, 'name': ReconfigVM_Task, 'duration_secs': 0.367592} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.429149] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037/83369251-b00f-4595-bf98-28a3ec84f037.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.430272] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f3f985-d49b-4e0b-a4c3-fa6c4b80bcaa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.435677] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df38685-75a4-4d02-840e-25d04c1d023a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.455333] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38a71d5-72c7-4f0a-a68c-e0c11f25e8df {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.465072] env[61995]: DEBUG nova.compute.provider_tree [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.484113] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839d348c-70d8-4fa6-ba69-f5f503c42b2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.505414] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232fd6ad-3695-4f7c-a941-a227b7943a94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.512343] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1129.512633] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfe47393-30dd-4f42-885c-bb971743d6db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.518611] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1129.518611] env[61995]: value = "task-795354" [ 1129.518611] env[61995]: _type = "Task" [ 1129.518611] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.527448] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795354, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.716158] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "d2f84704-7d96-4ed6-835f-7176ff10148a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.716466] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.716682] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "d2f84704-7d96-4ed6-835f-7176ff10148a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.716873] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.717062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.720225] env[61995]: INFO nova.compute.manager [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Terminating instance [ 1129.724078] env[61995]: DEBUG nova.compute.manager [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1129.724339] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1129.725247] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7196cf8-99b6-438d-9eb0-780a23c2e760 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.735402] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1129.735653] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3769dd95-2fd3-4a8d-a2af-8303c72e040c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.742665] env[61995]: DEBUG oslo_vmware.api [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1129.742665] env[61995]: value = "task-795355" [ 1129.742665] env[61995]: _type = "Task" [ 1129.742665] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.750242] env[61995]: DEBUG oslo_vmware.api [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.788952] env[61995]: DEBUG nova.compute.manager [req-c2dc6c47-3450-4bc4-9963-90eead136619 req-c1446952-b9d9-4967-9b76-7828c51dd6e5 service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Received event network-vif-deleted-011abc87-f621-435e-a18c-64d20e32d5f6 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1129.789103] env[61995]: INFO nova.compute.manager [req-c2dc6c47-3450-4bc4-9963-90eead136619 req-c1446952-b9d9-4967-9b76-7828c51dd6e5 service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Neutron deleted interface 011abc87-f621-435e-a18c-64d20e32d5f6; detaching it from the instance and deleting it from the info cache [ 1129.789339] env[61995]: DEBUG nova.network.neutron [req-c2dc6c47-3450-4bc4-9963-90eead136619 req-c1446952-b9d9-4967-9b76-7828c51dd6e5 service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.968761] env[61995]: DEBUG nova.scheduler.client.report [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.028973] env[61995]: DEBUG oslo_vmware.api [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795354, 'name': PowerOnVM_Task, 'duration_secs': 0.43203} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.028973] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1130.103110] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.103465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.103724] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.103965] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.104203] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.106596] env[61995]: INFO nova.compute.manager [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Terminating instance [ 1130.108691] env[61995]: DEBUG nova.compute.manager [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1130.108938] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1130.109881] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97b3728-c7b4-4dd3-99a7-4f6df0901d8f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.118567] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1130.118821] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f38fca3c-671a-4019-a2c2-eae1b598473f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.124868] env[61995]: DEBUG oslo_vmware.api [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1130.124868] env[61995]: value = "task-795356" [ 1130.124868] env[61995]: _type = "Task" [ 1130.124868] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.132659] env[61995]: DEBUG oslo_vmware.api [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795356, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.230951] env[61995]: DEBUG nova.network.neutron [-] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.252656] env[61995]: DEBUG oslo_vmware.api [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795355, 'name': PowerOffVM_Task, 'duration_secs': 0.178651} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.252916] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1130.253107] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1130.253361] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02043b2a-4d6f-49df-b977-e45b24890068 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.292712] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-440cf463-2d16-46ff-bf32-2cbe52ca0cac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.302118] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c75ce2b-4c37-4ac6-a98e-72f11dd21974 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.323121] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1130.323483] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1130.323655] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleting the datastore file [datastore1] d2f84704-7d96-4ed6-835f-7176ff10148a {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.323916] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c667167-d412-4fa3-be06-d4181a9d4f24 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.335405] env[61995]: DEBUG nova.compute.manager [req-c2dc6c47-3450-4bc4-9963-90eead136619 req-c1446952-b9d9-4967-9b76-7828c51dd6e5 service nova] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Detach interface failed, port_id=011abc87-f621-435e-a18c-64d20e32d5f6, reason: Instance 966783cb-b826-4983-9b77-32ab4ba56728 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1130.336869] env[61995]: DEBUG oslo_vmware.api [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1130.336869] env[61995]: value = "task-795358" [ 1130.336869] env[61995]: _type = "Task" [ 1130.336869] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.344297] env[61995]: DEBUG oslo_vmware.api [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.475287] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.477887] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.265s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.498468] env[61995]: INFO nova.scheduler.client.report [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted allocations for instance 50717fa4-956c-4e59-ab07-e6e8ec6f16bf [ 1130.508606] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "1142205a-bc9e-4a94-b201-2fc203f8f913" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.508761] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.636360] env[61995]: DEBUG oslo_vmware.api [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795356, 'name': PowerOffVM_Task, 'duration_secs': 0.2345} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.636516] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1130.636598] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1130.636859] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd376af3-31cc-4773-a2c0-9e6a2cfcce31 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.715140] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1130.715384] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1130.715573] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleting the datastore file [datastore1] 5cb4c1d9-e661-48b4-8600-1b823d63ddbb {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.715840] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c0028e75-045f-4ae1-9acf-7702e7ff184e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.722925] env[61995]: DEBUG oslo_vmware.api [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1130.722925] env[61995]: value = "task-795361" [ 1130.722925] env[61995]: _type = "Task" [ 1130.722925] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.730606] env[61995]: DEBUG oslo_vmware.api [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795361, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.734338] env[61995]: INFO nova.compute.manager [-] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Took 1.55 seconds to deallocate network for instance. [ 1130.846199] env[61995]: DEBUG oslo_vmware.api [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795358, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154916} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.846491] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.846700] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1130.846924] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1130.847100] env[61995]: INFO nova.compute.manager [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1130.847373] env[61995]: DEBUG oslo.service.loopingcall [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.847587] env[61995]: DEBUG nova.compute.manager [-] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1130.847687] env[61995]: DEBUG nova.network.neutron [-] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1130.983133] env[61995]: INFO nova.compute.claims [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1131.011404] env[61995]: DEBUG nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1131.013908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b45a5b06-895b-4249-8fe2-c0a319e4fd08 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "50717fa4-956c-4e59-ab07-e6e8ec6f16bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.456s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.079478] env[61995]: INFO nova.compute.manager [None req-ce3db27f-659d-415b-8cbe-100aaae347d4 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance to original state: 'active' [ 1131.233241] env[61995]: DEBUG oslo_vmware.api [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795361, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215156} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.233431] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.233648] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1131.233837] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1131.234076] env[61995]: INFO nova.compute.manager [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1131.234392] env[61995]: DEBUG oslo.service.loopingcall [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.234596] env[61995]: DEBUG nova.compute.manager [-] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1131.234712] env[61995]: DEBUG nova.network.neutron [-] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1131.240668] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.437194] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.437490] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.437709] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.437899] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.438138] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.440774] env[61995]: INFO nova.compute.manager [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Terminating instance [ 1131.442777] env[61995]: DEBUG nova.compute.manager [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1131.442983] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1131.443803] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e65f36c-b7a0-45e9-950d-31f1e9899416 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.452177] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1131.452436] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ccd350f-ecf5-4a9d-92b3-685f7237c8c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.459548] env[61995]: DEBUG oslo_vmware.api [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1131.459548] env[61995]: value = "task-795362" [ 1131.459548] env[61995]: _type = "Task" [ 1131.459548] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.467591] env[61995]: DEBUG oslo_vmware.api [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.489822] env[61995]: INFO nova.compute.resource_tracker [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating resource usage from migration 4c60573a-8da0-48d9-adde-dc3dacdd14d3 [ 1131.529316] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.587469] env[61995]: DEBUG nova.network.neutron [-] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.617592] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5ec5d6-fd44-496a-bd85-33c9179f8bb9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.627191] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43904d15-f4af-4e1c-98a1-c838aaae3221 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.661319] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8166e2a-427a-42e6-bdb3-1f648fdb825b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.668803] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b5e7c9-b6ff-458d-a4d1-a0fec1c221ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.683246] env[61995]: DEBUG nova.compute.provider_tree [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1131.818666] env[61995]: DEBUG nova.compute.manager [req-17c131b5-1bf6-4471-ab98-1c5c23e6974d req-96d08eda-44a9-4579-ab33-6065b42d0c06 service nova] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Received event network-vif-deleted-e697de59-3c5b-4928-a1df-9bf03e753d3c {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.818666] env[61995]: DEBUG nova.compute.manager [req-17c131b5-1bf6-4471-ab98-1c5c23e6974d req-96d08eda-44a9-4579-ab33-6065b42d0c06 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Received event network-vif-deleted-d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.818666] env[61995]: INFO nova.compute.manager [req-17c131b5-1bf6-4471-ab98-1c5c23e6974d req-96d08eda-44a9-4579-ab33-6065b42d0c06 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Neutron deleted interface d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9; detaching it from the instance and deleting it from the info cache [ 1131.818666] env[61995]: DEBUG nova.network.neutron [req-17c131b5-1bf6-4471-ab98-1c5c23e6974d req-96d08eda-44a9-4579-ab33-6065b42d0c06 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.974396] env[61995]: DEBUG oslo_vmware.api [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795362, 'name': PowerOffVM_Task, 'duration_secs': 0.354374} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.974859] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1131.976221] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1131.976221] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6da1b224-60ea-48f8-967f-0cf4066f37eb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.994821] env[61995]: DEBUG nova.network.neutron [-] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.085348] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1132.085485] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1132.085676] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleting the datastore file [datastore1] c3b2e000-cfa8-4b5b-8763-982c6b768621 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1132.085944] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d90dad0-ff51-4af5-b79e-ca684bec5928 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.090528] env[61995]: INFO nova.compute.manager [-] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Took 1.24 seconds to deallocate network for instance. [ 1132.094840] env[61995]: DEBUG oslo_vmware.api [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for the task: (returnval){ [ 1132.094840] env[61995]: value = "task-795365" [ 1132.094840] env[61995]: _type = "Task" [ 1132.094840] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.106332] env[61995]: DEBUG oslo_vmware.api [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795365, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.203498] env[61995]: ERROR nova.scheduler.client.report [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [req-dc6e82e3-3160-46d4-a2c8-37e30aa7acfb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-dc6e82e3-3160-46d4-a2c8-37e30aa7acfb"}]} [ 1132.219678] env[61995]: DEBUG nova.scheduler.client.report [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1132.234575] env[61995]: DEBUG nova.scheduler.client.report [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1132.234825] env[61995]: DEBUG nova.compute.provider_tree [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1132.248557] env[61995]: DEBUG nova.scheduler.client.report [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1132.267511] env[61995]: DEBUG nova.scheduler.client.report [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1132.321624] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef0664e5-3466-486c-a4ae-a8b1efbd123c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.332589] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea0e224-792b-4ed7-92d1-4baaf75042c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.360637] env[61995]: DEBUG nova.compute.manager [req-17c131b5-1bf6-4471-ab98-1c5c23e6974d req-96d08eda-44a9-4579-ab33-6065b42d0c06 service nova] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Detach interface failed, port_id=d8ff3c80-bfcb-4cdd-b0d2-96be0b176ab9, reason: Instance 5cb4c1d9-e661-48b4-8600-1b823d63ddbb could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1132.410957] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b2d247-dc1c-406c-8c51-cb8a7328736e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.418556] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930535b4-4f8e-475d-baea-55f8fbcf6456 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.450596] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bd3888-b33e-4035-8723-3182725e02c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.457822] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f5af68-411f-49a2-9434-7e9ed9f5f444 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.471135] env[61995]: DEBUG nova.compute.provider_tree [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1132.475152] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "83369251-b00f-4595-bf98-28a3ec84f037" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.475401] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.475773] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "83369251-b00f-4595-bf98-28a3ec84f037-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.475855] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.475960] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.479297] env[61995]: INFO nova.compute.manager [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Terminating instance [ 1132.481596] env[61995]: DEBUG nova.compute.manager [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1132.481813] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1132.482054] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba608962-329b-4ebc-872a-2ed36594df6c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.489300] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1132.489300] env[61995]: value = "task-795366" [ 1132.489300] env[61995]: _type = "Task" [ 1132.489300] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.497774] env[61995]: INFO nova.compute.manager [-] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Took 1.26 seconds to deallocate network for instance. [ 1132.498156] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795366, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.602695] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.608244] env[61995]: DEBUG oslo_vmware.api [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Task: {'id': task-795365, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184106} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.608498] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.608688] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1132.608865] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1132.609057] env[61995]: INFO nova.compute.manager [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1132.609356] env[61995]: DEBUG oslo.service.loopingcall [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.609565] env[61995]: DEBUG nova.compute.manager [-] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1132.609663] env[61995]: DEBUG nova.network.neutron [-] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1132.663826] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1132.664146] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185428', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'name': 'volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'serial': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1132.665147] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af0ee68-4552-4688-8f38-35da107385e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.683027] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67302df-3832-4f68-8107-f5788fa43ce5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.711827] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc/volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1132.714167] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a196df7-6fb8-4dd2-903a-8f5ba0630d9f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.733016] env[61995]: DEBUG oslo_vmware.api [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1132.733016] env[61995]: value = "task-795367" [ 1132.733016] env[61995]: _type = "Task" [ 1132.733016] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.741195] env[61995]: DEBUG oslo_vmware.api [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.999686] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795366, 'name': PowerOffVM_Task, 'duration_secs': 0.21092} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.000012] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1133.000229] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1133.000434] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185418', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'name': 'volume-ef951e09-0383-4bd5-891c-6e7ead61b641', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '83369251-b00f-4595-bf98-28a3ec84f037', 'attached_at': '2024-09-10T09:02:56.000000', 'detached_at': '', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'serial': 'ef951e09-0383-4bd5-891c-6e7ead61b641'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1133.001277] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c5164d-a81f-42fc-aaf7-a1d99685e3c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.005888] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.007365] env[61995]: DEBUG nova.scheduler.client.report [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 144 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1133.007365] env[61995]: DEBUG nova.compute.provider_tree [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 144 to 145 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1133.007365] env[61995]: DEBUG nova.compute.provider_tree [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1133.028551] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db59563f-a848-44de-8308-ea7fec812a65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.036045] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2ed4f6-c41e-43d9-8710-c7e971562688 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.057115] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b241b2cd-02d5-45c5-bd4c-a2a968b595fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.073809] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] The volume has not been displaced from its original location: [datastore2] volume-ef951e09-0383-4bd5-891c-6e7ead61b641/volume-ef951e09-0383-4bd5-891c-6e7ead61b641.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1133.079424] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.079743] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e1ad39a-170a-4a1e-ba90-e2e998cc9fa5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.096748] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1133.096748] env[61995]: value = "task-795368" [ 1133.096748] env[61995]: _type = "Task" [ 1133.096748] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.106892] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795368, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.242760] env[61995]: DEBUG oslo_vmware.api [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795367, 'name': ReconfigVM_Task, 'duration_secs': 0.363813} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.243095] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfigured VM instance instance-00000063 to attach disk [datastore2] volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc/volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1133.248335] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bff117a-f563-459f-8752-282c49db6edd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.267417] env[61995]: DEBUG oslo_vmware.api [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1133.267417] env[61995]: value = "task-795369" [ 1133.267417] env[61995]: _type = "Task" [ 1133.267417] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.282539] env[61995]: DEBUG oslo_vmware.api [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795369, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.408837] env[61995]: DEBUG nova.network.neutron [-] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.512417] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.034s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.512683] env[61995]: INFO nova.compute.manager [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Migrating [ 1133.519192] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.279s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.519830] env[61995]: DEBUG nova.objects.instance [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lazy-loading 'resources' on Instance uuid 966783cb-b826-4983-9b77-32ab4ba56728 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.607787] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795368, 'name': ReconfigVM_Task, 'duration_secs': 0.215356} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.608170] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1133.612593] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fda8837-a10d-467a-acc3-e563d757fa2d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.627011] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1133.627011] env[61995]: value = "task-795370" [ 1133.627011] env[61995]: _type = "Task" [ 1133.627011] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.634494] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.777103] env[61995]: DEBUG oslo_vmware.api [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795369, 'name': ReconfigVM_Task, 'duration_secs': 0.154265} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.777398] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185428', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'name': 'volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'serial': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1133.840405] env[61995]: DEBUG nova.compute.manager [req-fafcc1e2-92c3-409e-ae6c-fe8a2c1d8409 req-db0f4d33-1af1-43b9-91fe-97bb9846068a service nova] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Received event network-vif-deleted-7efb2852-0c7e-4e86-8996-8cba798b76b2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.911574] env[61995]: INFO nova.compute.manager [-] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Took 1.30 seconds to deallocate network for instance. [ 1134.028383] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.028661] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.028786] env[61995]: DEBUG nova.network.neutron [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1134.136523] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795370, 'name': ReconfigVM_Task, 'duration_secs': 0.129114} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.138942] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185418', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'name': 'volume-ef951e09-0383-4bd5-891c-6e7ead61b641', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '83369251-b00f-4595-bf98-28a3ec84f037', 'attached_at': '2024-09-10T09:02:56.000000', 'detached_at': '', 'volume_id': 'ef951e09-0383-4bd5-891c-6e7ead61b641', 'serial': 'ef951e09-0383-4bd5-891c-6e7ead61b641'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1134.139330] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1134.140550] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7bfcb8-618a-4e3b-afd6-de9636c00da4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.146606] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1134.146840] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-929da16f-32ef-4a41-bee5-5f4391bc94c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.160038] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1702d9-ec24-4884-bb47-bcd3fefb43b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.166969] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a93fb9-32bc-4259-92c1-c58e2e1f1a77 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.199773] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecee6f29-0014-4cf8-a23b-1a9ca1cfca1a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.207570] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42b31f3-f860-4fbe-b0eb-ca9f28d2a8b7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.222580] env[61995]: DEBUG nova.compute.provider_tree [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.227383] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1134.227560] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1134.228535] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleting the datastore file [datastore1] 83369251-b00f-4595-bf98-28a3ec84f037 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1134.228535] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7e0aa4c-f67c-4c96-8b04-5981d50a1068 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.233933] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1134.233933] env[61995]: value = "task-795372" [ 1134.233933] env[61995]: _type = "Task" [ 1134.233933] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.242512] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795372, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.417740] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.725703] env[61995]: DEBUG nova.scheduler.client.report [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.743699] env[61995]: DEBUG oslo_vmware.api [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795372, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239738} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.744551] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.744750] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1134.744933] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1134.745134] env[61995]: INFO nova.compute.manager [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Took 2.26 seconds to destroy the instance on the hypervisor. [ 1134.745378] env[61995]: DEBUG oslo.service.loopingcall [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.745569] env[61995]: DEBUG nova.compute.manager [-] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1134.745664] env[61995]: DEBUG nova.network.neutron [-] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1134.756652] env[61995]: DEBUG nova.network.neutron [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.811881] env[61995]: DEBUG nova.objects.instance [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid a73742da-3d5c-4b71-8e92-32e26d404f37 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.203208] env[61995]: DEBUG nova.compute.manager [req-44d58d81-cf5a-4cac-acd3-4efcdcaf2db7 req-efbefb57-0434-4b7e-b76e-03785139cd8f service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Received event network-vif-deleted-c97b92c2-09c6-42ae-bac7-ad1583a82097 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1135.203208] env[61995]: INFO nova.compute.manager [req-44d58d81-cf5a-4cac-acd3-4efcdcaf2db7 req-efbefb57-0434-4b7e-b76e-03785139cd8f service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Neutron deleted interface c97b92c2-09c6-42ae-bac7-ad1583a82097; detaching it from the instance and deleting it from the info cache [ 1135.203208] env[61995]: DEBUG nova.network.neutron [req-44d58d81-cf5a-4cac-acd3-4efcdcaf2db7 req-efbefb57-0434-4b7e-b76e-03785139cd8f service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.230821] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.234711] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.704s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.237882] env[61995]: INFO nova.compute.claims [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1135.261709] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.265895] env[61995]: INFO nova.scheduler.client.report [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Deleted allocations for instance 966783cb-b826-4983-9b77-32ab4ba56728 [ 1135.316221] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ce7d4cb2-d311-4ea8-bfca-9a9fe42505eb tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.255s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.595503] env[61995]: DEBUG oslo_concurrency.lockutils [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.595780] env[61995]: DEBUG oslo_concurrency.lockutils [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.681815] env[61995]: DEBUG nova.network.neutron [-] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.706299] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e945ebb6-2738-4eb7-abac-c6b12b1b8952 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.716322] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d5cfc4-35d3-48a9-b711-0dae2a6f30fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.745659] env[61995]: DEBUG nova.compute.manager [req-44d58d81-cf5a-4cac-acd3-4efcdcaf2db7 req-efbefb57-0434-4b7e-b76e-03785139cd8f service nova] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Detach interface failed, port_id=c97b92c2-09c6-42ae-bac7-ad1583a82097, reason: Instance 83369251-b00f-4595-bf98-28a3ec84f037 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1135.780584] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e3b31b0f-e8e8-4def-bc4d-b55a03b580b4 tempest-ServerMetadataTestJSON-908999434 tempest-ServerMetadataTestJSON-908999434-project-member] Lock "966783cb-b826-4983-9b77-32ab4ba56728" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.775s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.098793] env[61995]: INFO nova.compute.manager [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Detaching volume b7931f01-14ff-4e85-8499-8a899a9331a5 [ 1136.136841] env[61995]: INFO nova.virt.block_device [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Attempting to driver detach volume b7931f01-14ff-4e85-8499-8a899a9331a5 from mountpoint /dev/sdb [ 1136.136994] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1136.137214] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185424', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'name': 'volume-b7931f01-14ff-4e85-8499-8a899a9331a5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'serial': 'b7931f01-14ff-4e85-8499-8a899a9331a5'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1136.138173] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2059af-6792-4908-a650-e5affb17694a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.164308] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f502e44b-143c-47c1-9e99-fc88610b561b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.170664] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb81ace-cae8-49a1-9f20-d9850f053435 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.194208] env[61995]: INFO nova.compute.manager [-] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Took 1.45 seconds to deallocate network for instance. [ 1136.197749] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e5a8fb-4cb2-473b-bc29-b1ae70e9b004 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.214291] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] The volume has not been displaced from its original location: [datastore2] volume-b7931f01-14ff-4e85-8499-8a899a9331a5/volume-b7931f01-14ff-4e85-8499-8a899a9331a5.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1136.219791] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1136.220117] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c0c7be8-b3df-44af-9cca-ed81b62fa391 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.238143] env[61995]: DEBUG oslo_vmware.api [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1136.238143] env[61995]: value = "task-795373" [ 1136.238143] env[61995]: _type = "Task" [ 1136.238143] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.251571] env[61995]: DEBUG oslo_vmware.api [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795373, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.386187] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba2ab1a-fe07-43c1-a6ae-da4188892aea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.394607] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7451d5a-a993-4193-85fa-1a2382f5d8e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.427286] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58389896-f108-4a6c-bda1-9bb0fd3bb3da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.435734] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3f613f-6caa-4b21-b50f-1684ee2a6325 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.448947] env[61995]: DEBUG nova.compute.provider_tree [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.749063] env[61995]: DEBUG oslo_vmware.api [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795373, 'name': ReconfigVM_Task, 'duration_secs': 0.216513} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.749851] env[61995]: INFO nova.compute.manager [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Took 0.56 seconds to detach 1 volumes for instance. [ 1136.756942] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1136.761015] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41fb08bd-95f7-4147-8796-3fb06cb6445d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.782807] env[61995]: DEBUG oslo_vmware.api [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1136.782807] env[61995]: value = "task-795374" [ 1136.782807] env[61995]: _type = "Task" [ 1136.782807] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.783877] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e20cdd-ebdc-448a-8854-2b4ab7a9d120 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.795531] env[61995]: DEBUG oslo_vmware.api [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.809634] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance 'b2aa6712-3284-4d97-ac98-ff8789fa8bca' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1136.952883] env[61995]: DEBUG nova.scheduler.client.report [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1137.256880] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.298832] env[61995]: DEBUG oslo_vmware.api [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795374, 'name': ReconfigVM_Task, 'duration_secs': 0.144476} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.299201] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185424', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'name': 'volume-b7931f01-14ff-4e85-8499-8a899a9331a5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7931f01-14ff-4e85-8499-8a899a9331a5', 'serial': 'b7931f01-14ff-4e85-8499-8a899a9331a5'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1137.316262] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1137.317387] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bef224d5-8aad-4811-ae9a-c8775f20fe7d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.323497] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1137.323497] env[61995]: value = "task-795375" [ 1137.323497] env[61995]: _type = "Task" [ 1137.323497] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.332714] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795375, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.465290] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.465290] env[61995]: DEBUG nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1137.467602] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.865s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.469426] env[61995]: DEBUG nova.objects.instance [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lazy-loading 'resources' on Instance uuid d2f84704-7d96-4ed6-835f-7176ff10148a {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.835227] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795375, 'name': PowerOffVM_Task, 'duration_secs': 0.20218} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.835540] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1137.835540] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance 'b2aa6712-3284-4d97-ac98-ff8789fa8bca' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1137.846707] env[61995]: DEBUG nova.objects.instance [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid a73742da-3d5c-4b71-8e92-32e26d404f37 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.971165] env[61995]: DEBUG nova.compute.utils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1137.975314] env[61995]: DEBUG nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1137.975500] env[61995]: DEBUG nova.network.neutron [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1138.021566] env[61995]: DEBUG nova.policy [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '671c956f9796411e9b66268420075628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26c7cc2bdcb6462d8154127098e94875', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1138.088760] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0beb2ccc-2ac4-4224-87de-007bccee470c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.095148] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b25055-b94e-4df8-93cc-06c554761d03 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.125781] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7da2e0a-5ce6-493b-a158-5d8277ca9517 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.133255] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cc5a7a-d622-43dd-837e-e5eb1ca918a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.146464] env[61995]: DEBUG nova.compute.provider_tree [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.323972] env[61995]: DEBUG nova.network.neutron [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Successfully created port: f2a3f297-cb0b-4ba2-b472-65d810d6df37 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1138.341806] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1138.342082] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1138.342255] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1138.342449] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1138.342603] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1138.342757] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1138.342961] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1138.344142] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1138.344366] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1138.344548] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1138.344736] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1138.360136] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-870f200a-6efe-4cd9-8ef5-a41078ddcf43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.378070] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1138.378070] env[61995]: value = "task-795376" [ 1138.378070] env[61995]: _type = "Task" [ 1138.378070] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.387618] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795376, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.475601] env[61995]: DEBUG nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1138.653227] env[61995]: DEBUG nova.scheduler.client.report [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.861442] env[61995]: DEBUG oslo_concurrency.lockutils [None req-570fe0af-8867-4dd8-9c92-a89930125c5e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.265s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.887896] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795376, 'name': ReconfigVM_Task, 'duration_secs': 0.292405} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.888376] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance 'b2aa6712-3284-4d97-ac98-ff8789fa8bca' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1138.903726] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.903980] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.160202] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.166065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.159s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.166065] env[61995]: DEBUG nova.objects.instance [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lazy-loading 'resources' on Instance uuid 5cb4c1d9-e661-48b4-8600-1b823d63ddbb {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.188999] env[61995]: INFO nova.scheduler.client.report [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted allocations for instance d2f84704-7d96-4ed6-835f-7176ff10148a [ 1139.396438] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.396790] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.397014] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.397229] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.397386] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.397539] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.397747] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.397917] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.398143] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.398309] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.398495] env[61995]: DEBUG nova.virt.hardware [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.403691] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1139.405422] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-277d9f09-3f37-4645-bffa-d48ad630479e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.420620] env[61995]: INFO nova.compute.manager [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Detaching volume 17f21f6a-4284-4d68-b7ee-6d82e0d56cdc [ 1139.422588] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.422811] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.432034] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1139.432034] env[61995]: value = "task-795377" [ 1139.432034] env[61995]: _type = "Task" [ 1139.432034] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.441992] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795377, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.458234] env[61995]: INFO nova.virt.block_device [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Attempting to driver detach volume 17f21f6a-4284-4d68-b7ee-6d82e0d56cdc from mountpoint /dev/sdc [ 1139.458497] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1139.458691] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185428', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'name': 'volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'serial': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1139.459867] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bff01c9-023f-4c02-be58-56f4b1082591 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.483120] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae3e6cc-79bc-4533-abb0-71e4be5ff772 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.486405] env[61995]: DEBUG nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1139.493873] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb14f6a-b8f1-474a-b1de-adf40b6cb44b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.519249] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637c7be1-db93-4b59-8b51-95d0a5b316f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.535986] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] The volume has not been displaced from its original location: [datastore2] volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc/volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1139.541534] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfiguring VM instance instance-00000063 to detach disk 2002 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1139.543848] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.544100] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.544269] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.544687] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.544687] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.544805] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.544964] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.545147] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.545324] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.545494] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.545680] env[61995]: DEBUG nova.virt.hardware [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.545938] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84882e41-9e0f-4e1d-8c40-e00427824332 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.559683] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0671ca9a-73b3-400c-b9f8-56c9c1acc868 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.569052] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5482c68d-dd4e-4716-a8d4-d4afc1abb84f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.571792] env[61995]: DEBUG oslo_vmware.api [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1139.571792] env[61995]: value = "task-795378" [ 1139.571792] env[61995]: _type = "Task" [ 1139.571792] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.587082] env[61995]: DEBUG oslo_vmware.api [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795378, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.696831] env[61995]: DEBUG oslo_concurrency.lockutils [None req-dad44597-a535-4963-95d0-341b77f6168d tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "d2f84704-7d96-4ed6-835f-7176ff10148a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.980s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.806796] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0ec132-5905-4b6c-a418-6abda7e462e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.814561] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b9999a-f5c4-4a5e-9a13-96099c66fa9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.851043] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132951bf-2c8d-41c6-b355-f086427f51f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.859752] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef183e60-32fb-4b03-8360-b1791647f976 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.874086] env[61995]: DEBUG nova.compute.provider_tree [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.903867] env[61995]: DEBUG nova.compute.manager [req-910777ce-04b5-47d3-bf21-8aca01e107c1 req-1520c239-c8ad-49fb-9419-3c2ad3bc0a28 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Received event network-vif-plugged-f2a3f297-cb0b-4ba2-b472-65d810d6df37 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1139.904111] env[61995]: DEBUG oslo_concurrency.lockutils [req-910777ce-04b5-47d3-bf21-8aca01e107c1 req-1520c239-c8ad-49fb-9419-3c2ad3bc0a28 service nova] Acquiring lock "1142205a-bc9e-4a94-b201-2fc203f8f913-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.904335] env[61995]: DEBUG oslo_concurrency.lockutils [req-910777ce-04b5-47d3-bf21-8aca01e107c1 req-1520c239-c8ad-49fb-9419-3c2ad3bc0a28 service nova] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.904509] env[61995]: DEBUG oslo_concurrency.lockutils [req-910777ce-04b5-47d3-bf21-8aca01e107c1 req-1520c239-c8ad-49fb-9419-3c2ad3bc0a28 service nova] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.904682] env[61995]: DEBUG nova.compute.manager [req-910777ce-04b5-47d3-bf21-8aca01e107c1 req-1520c239-c8ad-49fb-9419-3c2ad3bc0a28 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] No waiting events found dispatching network-vif-plugged-f2a3f297-cb0b-4ba2-b472-65d810d6df37 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1139.904878] env[61995]: WARNING nova.compute.manager [req-910777ce-04b5-47d3-bf21-8aca01e107c1 req-1520c239-c8ad-49fb-9419-3c2ad3bc0a28 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Received unexpected event network-vif-plugged-f2a3f297-cb0b-4ba2-b472-65d810d6df37 for instance with vm_state building and task_state spawning. [ 1139.924934] env[61995]: DEBUG nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1139.941834] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795377, 'name': ReconfigVM_Task, 'duration_secs': 0.185609} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.942125] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1139.942918] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c36f9f-a8e1-491e-8cf8-3de404169e89 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.968818] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1139.969499] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08a66b15-1e79-4868-a31b-ba1da7bcfce1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.988974] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1139.988974] env[61995]: value = "task-795379" [ 1139.988974] env[61995]: _type = "Task" [ 1139.988974] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.997954] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795379, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.009539] env[61995]: DEBUG nova.network.neutron [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Successfully updated port: f2a3f297-cb0b-4ba2-b472-65d810d6df37 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1140.081962] env[61995]: DEBUG oslo_vmware.api [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795378, 'name': ReconfigVM_Task, 'duration_secs': 0.229788} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.082288] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Reconfigured VM instance instance-00000063 to detach disk 2002 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1140.087305] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaeeecfd-c393-4d6c-83f8-81e32acf1978 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.104032] env[61995]: DEBUG oslo_vmware.api [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1140.104032] env[61995]: value = "task-795380" [ 1140.104032] env[61995]: _type = "Task" [ 1140.104032] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.113145] env[61995]: DEBUG oslo_vmware.api [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.378286] env[61995]: DEBUG nova.scheduler.client.report [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.446469] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.499244] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795379, 'name': ReconfigVM_Task, 'duration_secs': 0.340235} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.499585] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfigured VM instance instance-00000067 to attach disk [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1140.499859] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance 'b2aa6712-3284-4d97-ac98-ff8789fa8bca' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1140.512043] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "refresh_cache-1142205a-bc9e-4a94-b201-2fc203f8f913" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.512179] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "refresh_cache-1142205a-bc9e-4a94-b201-2fc203f8f913" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.512321] env[61995]: DEBUG nova.network.neutron [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1140.595135] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.595353] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.615183] env[61995]: DEBUG oslo_vmware.api [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795380, 'name': ReconfigVM_Task, 'duration_secs': 0.141933} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.615514] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185428', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'name': 'volume-17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a73742da-3d5c-4b71-8e92-32e26d404f37', 'attached_at': '', 'detached_at': '', 'volume_id': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc', 'serial': '17f21f6a-4284-4d68-b7ee-6d82e0d56cdc'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1140.883491] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.885767] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.468s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.886018] env[61995]: DEBUG nova.objects.instance [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lazy-loading 'resources' on Instance uuid c3b2e000-cfa8-4b5b-8763-982c6b768621 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.906084] env[61995]: INFO nova.scheduler.client.report [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleted allocations for instance 5cb4c1d9-e661-48b4-8600-1b823d63ddbb [ 1141.007261] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74d4b79-21af-4f5b-8f32-3fdb96c20622 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.028018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e10eede-9c8e-4497-9cbd-062fe204e6d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.045827] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance 'b2aa6712-3284-4d97-ac98-ff8789fa8bca' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1141.057357] env[61995]: DEBUG nova.network.neutron [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1141.100668] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.100668] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1141.155456] env[61995]: DEBUG nova.objects.instance [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'flavor' on Instance uuid a73742da-3d5c-4b71-8e92-32e26d404f37 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.187110] env[61995]: DEBUG nova.network.neutron [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Updating instance_info_cache with network_info: [{"id": "f2a3f297-cb0b-4ba2-b472-65d810d6df37", "address": "fa:16:3e:4e:aa:41", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a3f297-cb", "ovs_interfaceid": "f2a3f297-cb0b-4ba2-b472-65d810d6df37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.413547] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f3c2e321-567f-4986-a0f9-517f16bd8bc0 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "5cb4c1d9-e661-48b4-8600-1b823d63ddbb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.310s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.491194] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961d9136-1db6-4059-9466-66a06abb4ae3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.498798] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d581640f-2dbc-4531-bc3a-fda9fc229999 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.528315] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904542c5-8712-4a59-8167-ca79fb4525a6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.535777] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8fb1ad-dc85-4e0c-aa06-6f49c2759505 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.548446] env[61995]: DEBUG nova.compute.provider_tree [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.608148] env[61995]: DEBUG nova.network.neutron [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Port 06334396-e435-480b-ab0c-0da719e15bf1 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1141.690162] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "refresh_cache-1142205a-bc9e-4a94-b201-2fc203f8f913" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.690475] env[61995]: DEBUG nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Instance network_info: |[{"id": "f2a3f297-cb0b-4ba2-b472-65d810d6df37", "address": "fa:16:3e:4e:aa:41", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a3f297-cb", "ovs_interfaceid": "f2a3f297-cb0b-4ba2-b472-65d810d6df37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1141.690887] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:aa:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2a3f297-cb0b-4ba2-b472-65d810d6df37', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1141.698505] env[61995]: DEBUG oslo.service.loopingcall [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1141.698715] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1141.698939] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d5f98d3-4f5b-43c9-a867-f59d5993fa18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.724619] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1141.724619] env[61995]: value = "task-795381" [ 1141.724619] env[61995]: _type = "Task" [ 1141.724619] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.733725] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795381, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.940352] env[61995]: DEBUG nova.compute.manager [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Received event network-changed-f2a3f297-cb0b-4ba2-b472-65d810d6df37 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1141.940600] env[61995]: DEBUG nova.compute.manager [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Refreshing instance network info cache due to event network-changed-f2a3f297-cb0b-4ba2-b472-65d810d6df37. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1141.940828] env[61995]: DEBUG oslo_concurrency.lockutils [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] Acquiring lock "refresh_cache-1142205a-bc9e-4a94-b201-2fc203f8f913" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.940973] env[61995]: DEBUG oslo_concurrency.lockutils [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] Acquired lock "refresh_cache-1142205a-bc9e-4a94-b201-2fc203f8f913" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.941320] env[61995]: DEBUG nova.network.neutron [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Refreshing network info cache for port f2a3f297-cb0b-4ba2-b472-65d810d6df37 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1142.052208] env[61995]: DEBUG nova.scheduler.client.report [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.163212] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ee424e82-2347-46e3-a599-00dabb1ff220 tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.259s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.219381] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "24b974c0-9567-46a3-96f6-fac1e602f46e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.219668] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.219886] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "24b974c0-9567-46a3-96f6-fac1e602f46e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.220476] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.220672] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.222773] env[61995]: INFO nova.compute.manager [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Terminating instance [ 1142.224469] env[61995]: DEBUG nova.compute.manager [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1142.224666] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1142.225535] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d94d11-96a4-43f0-962c-1a80c9872826 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.236857] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795381, 'name': CreateVM_Task, 'duration_secs': 0.331257} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.238794] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1142.239075] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1142.239665] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.239827] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.240160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1142.240374] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9826c49b-1330-4785-96a5-7163cd711bb1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.241880] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfd377da-7de1-4eea-8651-37bc2612f6e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.245936] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1142.245936] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5262b899-de61-55b9-2e05-0e04b79d92e2" [ 1142.245936] env[61995]: _type = "Task" [ 1142.245936] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.249936] env[61995]: DEBUG oslo_vmware.api [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1142.249936] env[61995]: value = "task-795382" [ 1142.249936] env[61995]: _type = "Task" [ 1142.249936] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.255452] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5262b899-de61-55b9-2e05-0e04b79d92e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.261455] env[61995]: DEBUG oslo_vmware.api [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.559007] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.561438] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.305s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.561648] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.563195] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.117s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.564998] env[61995]: INFO nova.compute.claims [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.630431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.630625] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.630809] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.664854] env[61995]: INFO nova.scheduler.client.report [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Deleted allocations for instance c3b2e000-cfa8-4b5b-8763-982c6b768621 [ 1142.666868] env[61995]: INFO nova.scheduler.client.report [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted allocations for instance 83369251-b00f-4595-bf98-28a3ec84f037 [ 1142.675344] env[61995]: DEBUG nova.network.neutron [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Updated VIF entry in instance network info cache for port f2a3f297-cb0b-4ba2-b472-65d810d6df37. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1142.676023] env[61995]: DEBUG nova.network.neutron [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Updating instance_info_cache with network_info: [{"id": "f2a3f297-cb0b-4ba2-b472-65d810d6df37", "address": "fa:16:3e:4e:aa:41", "network": {"id": "1644b82e-0a1c-4fcc-a46a-024ec85d6a51", "bridge": "br-int", "label": "tempest-ImagesTestJSON-842628016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26c7cc2bdcb6462d8154127098e94875", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a3f297-cb", "ovs_interfaceid": "f2a3f297-cb0b-4ba2-b472-65d810d6df37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.759573] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5262b899-de61-55b9-2e05-0e04b79d92e2, 'name': SearchDatastore_Task, 'duration_secs': 0.010078} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.760233] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.760484] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1142.760732] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.760885] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.761078] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1142.761329] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f3f2e48-072c-42f8-812c-07f8f782f9f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.766044] env[61995]: DEBUG oslo_vmware.api [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795382, 'name': PowerOffVM_Task, 'duration_secs': 0.203999} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.766674] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1142.766911] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1142.767106] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-866604ee-a118-4677-99e9-ddf456ed57fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.772600] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1142.772778] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1142.773451] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12d5b0e8-7f63-447e-aadf-4cc777f81438 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.777891] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1142.777891] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e2bdad-5c8b-6da7-d3d6-ee4d83206155" [ 1142.777891] env[61995]: _type = "Task" [ 1142.777891] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.784771] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e2bdad-5c8b-6da7-d3d6-ee4d83206155, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.837102] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1142.837338] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1142.837528] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleting the datastore file [datastore1] 24b974c0-9567-46a3-96f6-fac1e602f46e {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1142.837782] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19eb59af-d5c9-4f33-9308-65212fd9996b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.844445] env[61995]: DEBUG oslo_vmware.api [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for the task: (returnval){ [ 1142.844445] env[61995]: value = "task-795384" [ 1142.844445] env[61995]: _type = "Task" [ 1142.844445] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.851944] env[61995]: DEBUG oslo_vmware.api [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.176565] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b923e267-d6cb-441f-90e6-67ecee9d1d72 tempest-AttachInterfacesTestJSON-658626166 tempest-AttachInterfacesTestJSON-658626166-project-member] Lock "c3b2e000-cfa8-4b5b-8763-982c6b768621" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.739s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.178649] env[61995]: DEBUG oslo_concurrency.lockutils [req-b4a99391-f05b-4816-9578-62315327ad5f req-69c8489c-697c-4253-9c77-ce4faf3452a6 service nova] Releasing lock "refresh_cache-1142205a-bc9e-4a94-b201-2fc203f8f913" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.179510] env[61995]: DEBUG oslo_concurrency.lockutils [None req-122646b7-8c8e-4c74-a33d-e00ae87bb818 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "83369251-b00f-4595-bf98-28a3ec84f037" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.704s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.292297] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e2bdad-5c8b-6da7-d3d6-ee4d83206155, 'name': SearchDatastore_Task, 'duration_secs': 0.008173} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.293294] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39d39cab-47d5-44aa-9576-fb13f4faae10 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.306656] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1143.306656] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5291a543-18ed-2d0f-688b-1cce11c8d81b" [ 1143.306656] env[61995]: _type = "Task" [ 1143.306656] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.316434] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5291a543-18ed-2d0f-688b-1cce11c8d81b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.320278] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.320893] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.321408] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.321884] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.322352] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.326964] env[61995]: INFO nova.compute.manager [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Terminating instance [ 1143.329628] env[61995]: DEBUG nova.compute.manager [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1143.329963] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1143.331196] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9476339d-b4b1-40a5-b7eb-5a3306da1c80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.338893] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1143.339136] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5f89b53-a3ad-4958-bf4d-193f2271bc80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.345314] env[61995]: DEBUG oslo_vmware.api [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1143.345314] env[61995]: value = "task-795385" [ 1143.345314] env[61995]: _type = "Task" [ 1143.345314] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.356366] env[61995]: DEBUG oslo_vmware.api [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Task: {'id': task-795384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131508} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.359229] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.359479] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1143.359700] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1143.359915] env[61995]: INFO nova.compute.manager [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1143.360208] env[61995]: DEBUG oslo.service.loopingcall [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.360701] env[61995]: DEBUG oslo_vmware.api [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.360701] env[61995]: DEBUG nova.compute.manager [-] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1143.360818] env[61995]: DEBUG nova.network.neutron [-] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1143.676897] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.677164] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.678172] env[61995]: DEBUG nova.network.neutron [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1143.698009] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e0d5b4-2887-4176-885d-5af4963ea146 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.707626] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3d5b9b-9802-4cc0-952c-686a15bdbd66 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.737517] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2839dd90-8713-4c0f-8600-a690a1fbf607 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.745123] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4370bcd6-bf52-4770-8274-a38e80a7062e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.758245] env[61995]: DEBUG nova.compute.provider_tree [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.816068] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5291a543-18ed-2d0f-688b-1cce11c8d81b, 'name': SearchDatastore_Task, 'duration_secs': 0.04142} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.816401] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.816673] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 1142205a-bc9e-4a94-b201-2fc203f8f913/1142205a-bc9e-4a94-b201-2fc203f8f913.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1143.816935] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5e0881a-1cd6-4b39-a7f1-0b8542b28c57 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.823955] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1143.823955] env[61995]: value = "task-795386" [ 1143.823955] env[61995]: _type = "Task" [ 1143.823955] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.831887] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.857535] env[61995]: DEBUG oslo_vmware.api [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795385, 'name': PowerOffVM_Task, 'duration_secs': 0.195559} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.857805] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1143.857988] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1143.858322] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed258a44-b061-455f-88fd-e2ddc0408bd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.929985] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1143.930351] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1143.930625] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Deleting the datastore file [datastore2] a73742da-3d5c-4b71-8e92-32e26d404f37 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.930917] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a8b8ccf-3e1c-4579-a477-316eba8a5a81 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.939297] env[61995]: DEBUG oslo_vmware.api [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for the task: (returnval){ [ 1143.939297] env[61995]: value = "task-795388" [ 1143.939297] env[61995]: _type = "Task" [ 1143.939297] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.946890] env[61995]: DEBUG oslo_vmware.api [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.968234] env[61995]: DEBUG nova.compute.manager [req-30ca4450-3561-4fd3-aaa1-9033a60b52b4 req-daa69988-2326-4ac2-8d26-54b53eaa0b52 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Received event network-vif-deleted-2327f433-5c63-4d83-8050-6661d9ad60eb {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1143.968503] env[61995]: INFO nova.compute.manager [req-30ca4450-3561-4fd3-aaa1-9033a60b52b4 req-daa69988-2326-4ac2-8d26-54b53eaa0b52 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Neutron deleted interface 2327f433-5c63-4d83-8050-6661d9ad60eb; detaching it from the instance and deleting it from the info cache [ 1143.968743] env[61995]: DEBUG nova.network.neutron [req-30ca4450-3561-4fd3-aaa1-9033a60b52b4 req-daa69988-2326-4ac2-8d26-54b53eaa0b52 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.122232] env[61995]: DEBUG nova.network.neutron [-] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.176577] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.176994] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.261254] env[61995]: DEBUG nova.scheduler.client.report [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.337017] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795386, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.451128] env[61995]: DEBUG oslo_vmware.api [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.471343] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c459a5da-fc54-4a24-8b9e-f026ee2e3f38 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.480884] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ba1932-65ce-43ad-ba9b-6e75b9469383 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.501539] env[61995]: DEBUG nova.network.neutron [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.508991] env[61995]: DEBUG nova.compute.manager [req-30ca4450-3561-4fd3-aaa1-9033a60b52b4 req-daa69988-2326-4ac2-8d26-54b53eaa0b52 service nova] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Detach interface failed, port_id=2327f433-5c63-4d83-8050-6661d9ad60eb, reason: Instance 24b974c0-9567-46a3-96f6-fac1e602f46e could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1144.627375] env[61995]: INFO nova.compute.manager [-] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Took 1.27 seconds to deallocate network for instance. [ 1144.681713] env[61995]: DEBUG nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1144.767463] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.768057] env[61995]: DEBUG nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1144.835970] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795386, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690958} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.836330] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 1142205a-bc9e-4a94-b201-2fc203f8f913/1142205a-bc9e-4a94-b201-2fc203f8f913.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1144.836860] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1144.836860] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6cdd8720-34cb-4ace-a42c-7580dd5bb8a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.844771] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1144.844771] env[61995]: value = "task-795389" [ 1144.844771] env[61995]: _type = "Task" [ 1144.844771] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.852857] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.949939] env[61995]: DEBUG oslo_vmware.api [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Task: {'id': task-795388, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.784776} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.950444] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.953432] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1144.953432] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1144.953432] env[61995]: INFO nova.compute.manager [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1144.953432] env[61995]: DEBUG oslo.service.loopingcall [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1144.953432] env[61995]: DEBUG nova.compute.manager [-] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1144.953432] env[61995]: DEBUG nova.network.neutron [-] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1145.013380] env[61995]: DEBUG oslo_concurrency.lockutils [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.140019] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Didn't find any instances for network info cache update. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1145.140019] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.140019] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.140019] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.140019] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.140019] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.140019] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.140019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.140019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.141191] env[61995]: DEBUG nova.objects.instance [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lazy-loading 'resources' on Instance uuid 24b974c0-9567-46a3-96f6-fac1e602f46e {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.212479] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.273481] env[61995]: DEBUG nova.compute.utils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1145.275923] env[61995]: DEBUG nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Not allocating networking since 'none' was specified. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1145.357792] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063271} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.360613] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1145.361722] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9211ce56-abf1-49c3-b94f-ec4519bbc6c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.386876] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 1142205a-bc9e-4a94-b201-2fc203f8f913/1142205a-bc9e-4a94-b201-2fc203f8f913.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1145.387489] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f05e417c-458c-4875-b26f-9e11763a0507 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.407356] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1145.407356] env[61995]: value = "task-795390" [ 1145.407356] env[61995]: _type = "Task" [ 1145.407356] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.415681] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795390, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.543838] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3147c62-9d0d-452e-bbca-39dbf6b004f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.565244] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706f4949-5463-4369-a7da-c52a9f863b93 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.572996] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance 'b2aa6712-3284-4d97-ac98-ff8789fa8bca' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1145.644918] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Getting list of instances from cluster (obj){ [ 1145.644918] env[61995]: value = "domain-c8" [ 1145.644918] env[61995]: _type = "ClusterComputeResource" [ 1145.644918] env[61995]: } {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1145.649424] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022be242-c1e7-4ae2-a44d-1d4870f65a2b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.668592] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Got total of 2 instances {{(pid=61995) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1145.668804] env[61995]: WARNING nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] While synchronizing instance power states, found 5 instances in the database and 2 instances on the hypervisor. [ 1145.668978] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Triggering sync for uuid 24b974c0-9567-46a3-96f6-fac1e602f46e {{(pid=61995) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1145.669351] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Triggering sync for uuid a73742da-3d5c-4b71-8e92-32e26d404f37 {{(pid=61995) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1145.669638] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Triggering sync for uuid b2aa6712-3284-4d97-ac98-ff8789fa8bca {{(pid=61995) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1145.669837] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Triggering sync for uuid 1142205a-bc9e-4a94-b201-2fc203f8f913 {{(pid=61995) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1145.669996] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Triggering sync for uuid 1c2afe20-e28f-4553-a6e1-7f31ec0db64b {{(pid=61995) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1145.673180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "24b974c0-9567-46a3-96f6-fac1e602f46e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.673180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "a73742da-3d5c-4b71-8e92-32e26d404f37" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.673180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.673180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.673180] env[61995]: INFO nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] During sync_power_state the instance has a pending task (resize_finish). Skip. [ 1145.673180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.673180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "1142205a-bc9e-4a94-b201-2fc203f8f913" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.673180] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.673180] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.673180] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1145.673180] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.776670] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2961641a-e006-40c6-9db9-936c54d0e7ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.781333] env[61995]: DEBUG nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1145.787936] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9501fb3d-64b0-46e1-a5d9-30b67337969e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.821539] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e5c9be-3600-452d-b3eb-0c6d0d9a4b32 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.831936] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7419b60d-83d3-4973-84dd-35bf6fd94e15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.844994] env[61995]: DEBUG nova.compute.provider_tree [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.917644] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795390, 'name': ReconfigVM_Task, 'duration_secs': 0.33564} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.917946] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 1142205a-bc9e-4a94-b201-2fc203f8f913/1142205a-bc9e-4a94-b201-2fc203f8f913.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1145.918625] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e422e0c-a445-40c9-b541-dd0a1f7f8d4b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.926382] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1145.926382] env[61995]: value = "task-795391" [ 1145.926382] env[61995]: _type = "Task" [ 1145.926382] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.933823] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795391, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.996413] env[61995]: DEBUG nova.compute.manager [req-04807df1-1a0d-458b-bd4c-60348b72c9f8 req-154a42fc-81fa-44a7-9a4b-f310157555d9 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Received event network-vif-deleted-2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.996572] env[61995]: INFO nova.compute.manager [req-04807df1-1a0d-458b-bd4c-60348b72c9f8 req-154a42fc-81fa-44a7-9a4b-f310157555d9 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Neutron deleted interface 2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7; detaching it from the instance and deleting it from the info cache [ 1145.996692] env[61995]: DEBUG nova.network.neutron [req-04807df1-1a0d-458b-bd4c-60348b72c9f8 req-154a42fc-81fa-44a7-9a4b-f310157555d9 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.080617] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1146.080937] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbb8a60d-1264-4806-9747-620f156e8b5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.087889] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1146.087889] env[61995]: value = "task-795392" [ 1146.087889] env[61995]: _type = "Task" [ 1146.087889] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.097845] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.157565] env[61995]: DEBUG nova.network.neutron [-] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.176262] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.347788] env[61995]: DEBUG nova.scheduler.client.report [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.437154] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795391, 'name': Rename_Task, 'duration_secs': 0.152383} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.437469] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1146.437788] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff248b71-96c3-43d4-88c1-5914bfaf5b32 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.445051] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1146.445051] env[61995]: value = "task-795393" [ 1146.445051] env[61995]: _type = "Task" [ 1146.445051] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.452735] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795393, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.499295] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee30bb92-29ba-40d6-b6c2-d177e49ab537 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.508127] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46cc1b73-4e95-40b8-bd30-3c009c00bb91 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.533045] env[61995]: DEBUG nova.compute.manager [req-04807df1-1a0d-458b-bd4c-60348b72c9f8 req-154a42fc-81fa-44a7-9a4b-f310157555d9 service nova] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Detach interface failed, port_id=2ebf2f2d-139c-4be4-96b0-68a8ae90f0b7, reason: Instance a73742da-3d5c-4b71-8e92-32e26d404f37 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1146.599777] env[61995]: DEBUG oslo_vmware.api [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795392, 'name': PowerOnVM_Task, 'duration_secs': 0.398754} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.600280] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1146.600569] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-b4c61937-9724-4a44-883b-9bc498f0711c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance 'b2aa6712-3284-4d97-ac98-ff8789fa8bca' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1146.665723] env[61995]: INFO nova.compute.manager [-] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Took 1.71 seconds to deallocate network for instance. [ 1146.793541] env[61995]: DEBUG nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1146.826053] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.826053] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.826053] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.826053] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.826053] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.826053] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.826301] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.826478] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.826670] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.826847] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.827057] env[61995]: DEBUG nova.virt.hardware [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.828196] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67be8657-6ca3-4654-9c32-3226ac9227e1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.837281] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca16871-f062-416f-acda-3424ae9a00a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.853468] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.713s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.855665] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1146.861586] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Creating folder: Project (8252198f9ba643ba9e79b5a62713b57a). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1146.862930] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.650s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.863899] env[61995]: INFO nova.compute.claims [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1146.867271] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a2abcbb-14f7-49a0-a37d-bc9b1175227e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.879952] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Created folder: Project (8252198f9ba643ba9e79b5a62713b57a) in parent group-v185203. [ 1146.880350] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Creating folder: Instances. Parent ref: group-v185430. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1146.880716] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-236fe9d7-dbb4-4936-9e3a-35738166dc6b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.884533] env[61995]: INFO nova.scheduler.client.report [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Deleted allocations for instance 24b974c0-9567-46a3-96f6-fac1e602f46e [ 1146.892139] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Created folder: Instances in parent group-v185430. [ 1146.892395] env[61995]: DEBUG oslo.service.loopingcall [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.892593] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1146.892818] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d53a09a-be2f-4afd-8740-cfb66d6321b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.915527] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1146.915527] env[61995]: value = "task-795396" [ 1146.915527] env[61995]: _type = "Task" [ 1146.915527] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.923453] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795396, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.955064] env[61995]: DEBUG oslo_vmware.api [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795393, 'name': PowerOnVM_Task, 'duration_secs': 0.454058} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.955345] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1146.955550] env[61995]: INFO nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Took 7.47 seconds to spawn the instance on the hypervisor. [ 1146.955732] env[61995]: DEBUG nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1146.956531] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b2ee78-044b-444b-8185-7a4cafa4bdca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.174837] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.393177] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4a0d75d9-1238-42aa-9d14-a6dd0f9519d9 tempest-ServerRescueNegativeTestJSON-1356175332 tempest-ServerRescueNegativeTestJSON-1356175332-project-member] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.173s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.394596] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.724s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.394886] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc6347df-2b54-4130-a734-2cbdb4177d53 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.405360] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2dcf4f-6916-443a-b858-8290d21c5996 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.429822] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795396, 'name': CreateVM_Task, 'duration_secs': 0.402392} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.430034] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1147.430552] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.430709] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.431058] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1147.431313] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-700f6def-e1f7-4753-a425-a964554b6e46 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.442742] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1147.442742] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52444b06-ab99-072c-4333-b3a433fd7897" [ 1147.442742] env[61995]: _type = "Task" [ 1147.442742] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.458315] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52444b06-ab99-072c-4333-b3a433fd7897, 'name': SearchDatastore_Task, 'duration_secs': 0.010781} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.458315] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.458315] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1147.458315] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.458315] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.458315] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1147.458315] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a32ace6a-971e-48c1-b4c9-fd488b774283 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.470629] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1147.470818] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1147.473902] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24d49f98-7ea9-468c-928f-3c7dcbc1b798 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.477933] env[61995]: INFO nova.compute.manager [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Took 15.96 seconds to build instance. [ 1147.483895] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1147.483895] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5266f1f0-a983-6dbc-af0e-742d99d067d4" [ 1147.483895] env[61995]: _type = "Task" [ 1147.483895] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.496313] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5266f1f0-a983-6dbc-af0e-742d99d067d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.943414] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "24b974c0-9567-46a3-96f6-fac1e602f46e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.549s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.984564] env[61995]: DEBUG oslo_concurrency.lockutils [None req-36f4c4bc-c9c7-4b56-8286-21ea35bfde78 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.474s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.985256] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.313s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.985256] env[61995]: INFO nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] During sync_power_state the instance has a pending task (spawning). Skip. [ 1147.985673] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.002069] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5266f1f0-a983-6dbc-af0e-742d99d067d4, 'name': SearchDatastore_Task, 'duration_secs': 0.01926} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.002069] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bd7c26e-1d46-42d2-a7e2-9aa7eb46b651 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.005280] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1148.005280] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521a25d6-6c8a-4298-255b-bb313e2530c4" [ 1148.005280] env[61995]: _type = "Task" [ 1148.005280] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.020912] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521a25d6-6c8a-4298-255b-bb313e2530c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.037687] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed10720-d432-4932-97ae-17a4143791f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.046915] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce71b93-5da8-4bb8-92f6-1d89535158bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.084146] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f4887b-f7fb-4691-a0f6-e5341c6dcd24 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.092654] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43347030-7f44-4fb4-9c37-1cd7a251951f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.112787] env[61995]: DEBUG nova.compute.provider_tree [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.522229] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521a25d6-6c8a-4298-255b-bb313e2530c4, 'name': SearchDatastore_Task, 'duration_secs': 0.027521} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.522683] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.523139] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1148.523814] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a4c817b-85f5-4e92-9f1b-cb01d1aa80c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.534185] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1148.534185] env[61995]: value = "task-795397" [ 1148.534185] env[61995]: _type = "Task" [ 1148.534185] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.543291] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795397, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.615660] env[61995]: DEBUG nova.scheduler.client.report [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1148.717019] env[61995]: DEBUG nova.network.neutron [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Port 06334396-e435-480b-ab0c-0da719e15bf1 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1148.717019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.717019] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.717019] env[61995]: DEBUG nova.network.neutron [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1149.046117] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795397, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.122690] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.126111] env[61995]: DEBUG nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1149.129382] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.952s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.130023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.130023] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1149.130023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.955s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.130340] env[61995]: DEBUG nova.objects.instance [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lazy-loading 'resources' on Instance uuid a73742da-3d5c-4b71-8e92-32e26d404f37 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.132959] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde8f167-82eb-4d76-ad53-0c6c08dea461 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.142508] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8401e38e-6d92-48f0-b412-f617dadd8b99 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.161629] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b64a666-df02-4d5c-954d-200688dfe170 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.169296] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89332ff9-4422-490f-8b28-ef6df1bb8109 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.202924] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179740MB free_disk=67GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1149.203122] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.486357] env[61995]: DEBUG nova.compute.manager [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.487272] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91520e2-041f-4cfd-9558-389299c7d192 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.544975] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795397, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539288} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.544975] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1149.545478] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1149.545570] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71d4a68c-3aec-400c-8021-e9c5665c84b6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.552750] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1149.552750] env[61995]: value = "task-795398" [ 1149.552750] env[61995]: _type = "Task" [ 1149.552750] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.562965] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795398, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.573834] env[61995]: DEBUG nova.network.neutron [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.634690] env[61995]: DEBUG nova.compute.utils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1149.636672] env[61995]: DEBUG nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1149.637166] env[61995]: DEBUG nova.network.neutron [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1149.691419] env[61995]: DEBUG nova.policy [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4fa18e6b6fe742909fa15a846d3019e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd23f8abd8f14ec392fbfb7fd5bc64f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1149.698617] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "954d0ada-a870-415e-bd50-c066d27b7026" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.698859] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "954d0ada-a870-415e-bd50-c066d27b7026" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.765580] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69d4efc-1167-47d0-ab1b-79e80538dd08 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.774241] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1aa420-a626-4f80-afcf-5703bfb9a8bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.807020] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1251ee92-dfe9-41b7-b6a6-48618020c11a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.816891] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d038c91f-3bdd-4ca6-85b1-46d08ff1572b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.830503] env[61995]: DEBUG nova.compute.provider_tree [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.988081] env[61995]: DEBUG nova.network.neutron [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Successfully created port: fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1149.999097] env[61995]: INFO nova.compute.manager [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] instance snapshotting [ 1150.003452] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78b7d8f-f863-4c7c-a273-31d31cb88111 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.024589] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c500a0a-7f05-47a6-b578-87bac4da532c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.063473] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795398, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067782} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.063803] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1150.064535] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eeb10b1-68e6-4cb4-99e8-e506f7b99db1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.077745] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.088018] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.089247] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaef1ee2-5bbd-42c5-9070-d4c3dda4a757 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.116997] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1150.116997] env[61995]: value = "task-795399" [ 1150.116997] env[61995]: _type = "Task" [ 1150.116997] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.129810] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795399, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.140567] env[61995]: DEBUG nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1150.204980] env[61995]: DEBUG nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1150.333649] env[61995]: DEBUG nova.scheduler.client.report [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.538731] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1150.539086] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-352725fe-ec88-4972-80ef-e0998a0ece3a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.546706] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1150.546706] env[61995]: value = "task-795400" [ 1150.546706] env[61995]: _type = "Task" [ 1150.546706] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.555111] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795400, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.590352] env[61995]: DEBUG nova.compute.manager [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61995) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1150.590630] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.626817] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.729295] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.840044] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.842507] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.639s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.858996] env[61995]: INFO nova.scheduler.client.report [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Deleted allocations for instance a73742da-3d5c-4b71-8e92-32e26d404f37 [ 1151.056719] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795400, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.140968] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795399, 'name': ReconfigVM_Task, 'duration_secs': 0.662122} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.141424] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1151.142930] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c8ab02c-9dbe-4d5f-840b-725803f03947 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.160643] env[61995]: DEBUG nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1151.164812] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1151.164812] env[61995]: value = "task-795401" [ 1151.164812] env[61995]: _type = "Task" [ 1151.164812] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.182738] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795401, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.217137] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1151.217570] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1151.217902] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1151.218244] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1151.218467] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1151.218727] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1151.219045] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1151.219312] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1151.219571] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1151.220212] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1151.220212] env[61995]: DEBUG nova.virt.hardware [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1151.221393] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cc3221-30f0-4683-b97e-1c4bbec371e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.231013] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b135e22e-2898-40f1-9319-da497c2de8b9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.366862] env[61995]: DEBUG oslo_concurrency.lockutils [None req-92f42df2-0507-4de2-b257-2b775023798e tempest-AttachVolumeTestJSON-786439589 tempest-AttachVolumeTestJSON-786439589-project-member] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.046s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.367392] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.696s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.367593] env[61995]: INFO nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] During sync_power_state the instance has a pending task (deleting). Skip. [ 1151.367771] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "a73742da-3d5c-4b71-8e92-32e26d404f37" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.522883] env[61995]: DEBUG nova.compute.manager [req-f6fdd2a3-9907-4855-9dd3-ad560c09d6d6 req-bc7aa736-7f23-4c65-b245-e41ca368ff09 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-vif-plugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1151.523118] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6fdd2a3-9907-4855-9dd3-ad560c09d6d6 req-bc7aa736-7f23-4c65-b245-e41ca368ff09 service nova] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.523340] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6fdd2a3-9907-4855-9dd3-ad560c09d6d6 req-bc7aa736-7f23-4c65-b245-e41ca368ff09 service nova] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.523515] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6fdd2a3-9907-4855-9dd3-ad560c09d6d6 req-bc7aa736-7f23-4c65-b245-e41ca368ff09 service nova] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.523686] env[61995]: DEBUG nova.compute.manager [req-f6fdd2a3-9907-4855-9dd3-ad560c09d6d6 req-bc7aa736-7f23-4c65-b245-e41ca368ff09 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] No waiting events found dispatching network-vif-plugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1151.523857] env[61995]: WARNING nova.compute.manager [req-f6fdd2a3-9907-4855-9dd3-ad560c09d6d6 req-bc7aa736-7f23-4c65-b245-e41ca368ff09 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received unexpected event network-vif-plugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 for instance with vm_state building and task_state spawning. [ 1151.557770] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795400, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.628026] env[61995]: DEBUG nova.network.neutron [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Successfully updated port: fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1151.677562] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795401, 'name': Rename_Task, 'duration_secs': 0.142801} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.677769] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1151.678090] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54cd782d-29d6-4aee-824e-ba9254976c43 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.684492] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1151.684492] env[61995]: value = "task-795402" [ 1151.684492] env[61995]: _type = "Task" [ 1151.684492] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.693221] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795402, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.850541] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Applying migration context for instance b2aa6712-3284-4d97-ac98-ff8789fa8bca as it has an incoming, in-progress migration 4c60573a-8da0-48d9-adde-dc3dacdd14d3. Migration status is reverting {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1151.851353] env[61995]: INFO nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating resource usage from migration 4c60573a-8da0-48d9-adde-dc3dacdd14d3 [ 1151.870507] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Migration 4c60573a-8da0-48d9-adde-dc3dacdd14d3 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1151.870687] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance b2aa6712-3284-4d97-ac98-ff8789fa8bca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.870813] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 1142205a-bc9e-4a94-b201-2fc203f8f913 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.870933] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 1c2afe20-e28f-4553-a6e1-7f31ec0db64b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.871066] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance da5c39c9-f733-4452-9c7c-c92830682428 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.066016] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795400, 'name': CreateSnapshot_Task, 'duration_secs': 1.185089} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.066451] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1152.067245] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0b9439-95dc-48fc-bdf7-e054e146c3b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.130947] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.131148] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.131307] env[61995]: DEBUG nova.network.neutron [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1152.198021] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795402, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.374376] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 954d0ada-a870-415e-bd50-c066d27b7026 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1152.374576] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1152.374664] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1152.465566] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50bf7b1-83af-4996-8961-41c7f3e65d83 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.472693] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2baeb9c1-74c0-45d1-b3df-2174edb218f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.504053] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3244c967-7317-41bd-b7a0-f9e7b1c6056b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.511256] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8190a051-f215-47e5-ae94-e81e9ae8a5f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.524160] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.586258] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1152.586495] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d13acfb3-9e60-405b-9177-2fe4598590b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.595106] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1152.595106] env[61995]: value = "task-795404" [ 1152.595106] env[61995]: _type = "Task" [ 1152.595106] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.602572] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795404, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.662027] env[61995]: DEBUG nova.network.neutron [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1152.694809] env[61995]: DEBUG oslo_vmware.api [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795402, 'name': PowerOnVM_Task, 'duration_secs': 0.776577} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.695454] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1152.695902] env[61995]: INFO nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Took 5.90 seconds to spawn the instance on the hypervisor. [ 1152.696121] env[61995]: DEBUG nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.697295] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557f1e3b-c543-4365-8ce4-311579797bf5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.840016] env[61995]: DEBUG nova.network.neutron [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.026812] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1153.105052] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795404, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.216853] env[61995]: INFO nova.compute.manager [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Took 12.79 seconds to build instance. [ 1153.342921] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.343042] env[61995]: DEBUG nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance network_info: |[{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1153.343490] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:d3:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7894814c-6be3-4b80-a08e-4a771bc05dd1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fef53ce7-5db4-47e6-a55a-7da9bff3cda2', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1153.351299] env[61995]: DEBUG oslo.service.loopingcall [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1153.351533] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1153.351811] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-406130be-9bdf-4173-a4f0-6ebd7588dccd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.372680] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1153.372680] env[61995]: value = "task-795405" [ 1153.372680] env[61995]: _type = "Task" [ 1153.372680] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.380969] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795405, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.517364] env[61995]: INFO nova.compute.manager [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Rebuilding instance [ 1153.531417] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1153.531660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.689s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.534073] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.943s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.536314] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.536521] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Cleaning up deleted instances {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1153.550947] env[61995]: DEBUG nova.compute.manager [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1153.551308] env[61995]: DEBUG nova.compute.manager [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing instance network info cache due to event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1153.551463] env[61995]: DEBUG oslo_concurrency.lockutils [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.551663] env[61995]: DEBUG oslo_concurrency.lockutils [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.551785] env[61995]: DEBUG nova.network.neutron [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1153.568671] env[61995]: DEBUG nova.compute.manager [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1153.569599] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72e6169-edb1-4c73-83f0-d4fe0b1f9ef8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.607707] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795404, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.719087] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e7d9c343-b034-47dd-9b33-d6fbfd47ddf2 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.296s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.719900] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 8.047s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.719900] env[61995]: INFO nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] During sync_power_state the instance has a pending task (networking). Skip. [ 1153.719900] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.882681] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795405, 'name': CreateVM_Task, 'duration_secs': 0.355055} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.882859] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1153.883538] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.883712] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.884052] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1153.884320] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8db35405-62fd-436c-ae1f-f573e0a698d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.888922] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1153.888922] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a7217c-556e-8152-7cb1-5b850fa6224e" [ 1153.888922] env[61995]: _type = "Task" [ 1153.888922] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.896264] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a7217c-556e-8152-7cb1-5b850fa6224e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.040355] env[61995]: DEBUG nova.objects.instance [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'migration_context' on Instance uuid b2aa6712-3284-4d97-ac98-ff8789fa8bca {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.055703] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] There are 62 instances to clean {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1154.055871] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 966783cb-b826-4983-9b77-32ab4ba56728] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.084401] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1154.084699] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85160918-ad9d-4a39-a08b-f9ded65fb2c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.097377] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1154.097377] env[61995]: value = "task-795406" [ 1154.097377] env[61995]: _type = "Task" [ 1154.097377] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.112299] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795406, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.112552] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795404, 'name': CloneVM_Task} progress is 95%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.153423] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5162def-da3c-4c5c-bfdf-ae7d2a5fd674 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.163192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3794abf9-c2ea-42f7-aa71-5eb1be65e250 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.199519] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad76da7-13f5-4d37-9b90-2166d287212c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.207409] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06aed5f-65a0-4b94-b0ad-27d193ef15f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.221537] env[61995]: DEBUG nova.compute.provider_tree [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.302422] env[61995]: DEBUG nova.network.neutron [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updated VIF entry in instance network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.303275] env[61995]: DEBUG nova.network.neutron [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.399522] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a7217c-556e-8152-7cb1-5b850fa6224e, 'name': SearchDatastore_Task, 'duration_secs': 0.011827} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.399800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.400059] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1154.400310] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.400480] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.400686] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1154.400947] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de79f62d-4e33-40c3-b9f1-55451f071107 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.411751] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1154.411942] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1154.412646] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eabbc3f4-7506-4704-a8cc-314da88a2487 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.417565] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1154.417565] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]528cc068-5424-1b76-a372-9409dd144567" [ 1154.417565] env[61995]: _type = "Task" [ 1154.417565] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.424912] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528cc068-5424-1b76-a372-9409dd144567, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.564291] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 55a1c5d3-7e00-40f6-9b06-f206bf2bf895] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.609826] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795406, 'name': PowerOffVM_Task, 'duration_secs': 0.142707} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.612825] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1154.613065] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1154.613351] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795404, 'name': CloneVM_Task, 'duration_secs': 1.644367} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.614020] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60d61e0-85c0-41a6-98d8-f95727358921 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.616414] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Created linked-clone VM from snapshot [ 1154.617096] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f43252-a996-451f-8b88-15c2de18c83f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.624385] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Uploading image 5a9e44be-fb00-4d0c-b088-3150197a6c5a {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1154.628118] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1154.628360] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d114ef70-437b-4cad-a848-3d1ce6127bb9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.639026] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1154.639255] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a9bbf7b3-762f-4b94-a4c5-8d21a5fc1280 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.644626] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1154.644626] env[61995]: value = "task-795408" [ 1154.644626] env[61995]: _type = "Task" [ 1154.644626] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.652986] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795408, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.654486] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1154.654694] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1154.654873] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Deleting the datastore file [datastore2] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.655127] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c66c9ef-99da-4f29-b532-8c3a5ed06c61 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.662414] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1154.662414] env[61995]: value = "task-795409" [ 1154.662414] env[61995]: _type = "Task" [ 1154.662414] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.670161] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795409, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.724813] env[61995]: DEBUG nova.scheduler.client.report [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1154.805997] env[61995]: DEBUG oslo_concurrency.lockutils [req-4ea76184-21a1-4f7c-bd69-7479dd98c3d8 req-fc2ff966-8f4e-418f-bd75-79d7a2627ad1 service nova] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.931173] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]528cc068-5424-1b76-a372-9409dd144567, 'name': SearchDatastore_Task, 'duration_secs': 0.013673} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.931941] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ff7779a-06fc-47f2-a300-53678e15fbb5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.937217] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1154.937217] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529a21ff-80ec-d0fe-843e-eac29c8c0e3d" [ 1154.937217] env[61995]: _type = "Task" [ 1154.937217] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.944697] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529a21ff-80ec-d0fe-843e-eac29c8c0e3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.067307] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d2f84704-7d96-4ed6-835f-7176ff10148a] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.154510] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795408, 'name': Destroy_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.171628] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795409, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118446} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.171900] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.172125] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1155.172326] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1155.448034] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529a21ff-80ec-d0fe-843e-eac29c8c0e3d, 'name': SearchDatastore_Task, 'duration_secs': 0.008704} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.448437] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.448546] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1155.448810] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aeda2a74-f806-4cb7-9b7b-d165f69f503d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.456476] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1155.456476] env[61995]: value = "task-795411" [ 1155.456476] env[61995]: _type = "Task" [ 1155.456476] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.463997] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.572062] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 50717fa4-956c-4e59-ab07-e6e8ec6f16bf] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.655978] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795408, 'name': Destroy_Task, 'duration_secs': 0.625513} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.656264] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Destroyed the VM [ 1155.656536] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1155.656801] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ee328271-287c-4f5a-9d90-a6284a8ee96e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.663748] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1155.663748] env[61995]: value = "task-795412" [ 1155.663748] env[61995]: _type = "Task" [ 1155.663748] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.673749] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795412, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.735467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.201s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.742302] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.013s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.744053] env[61995]: INFO nova.compute.claims [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1155.967180] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795411, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45019} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.967456] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1155.967673] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1155.967920] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ef0ac32-80fb-450b-a6d5-f446551ac54f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.974291] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1155.974291] env[61995]: value = "task-795413" [ 1155.974291] env[61995]: _type = "Task" [ 1155.974291] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.981113] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795413, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.075262] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 85c56f36-a858-4b81-978d-bcd4f09d2ad8] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.174632] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795412, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.210176] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1156.210485] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1156.211078] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1156.211078] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1156.211344] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1156.211546] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1156.211808] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1156.212025] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1156.212263] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1156.212472] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1156.212670] env[61995]: DEBUG nova.virt.hardware [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.214081] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729d6f02-bfff-4409-b2ef-4379e93665e4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.221973] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e476a83f-abb5-4bf5-b5e6-e02448127803 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.235398] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Instance VIF info [] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1156.240918] env[61995]: DEBUG oslo.service.loopingcall [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.241203] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1156.241449] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fcfb6572-dd1d-4c6f-9c0e-8838d578787f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.261746] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1156.261746] env[61995]: value = "task-795414" [ 1156.261746] env[61995]: _type = "Task" [ 1156.261746] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.269941] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795414, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.483775] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795413, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0614} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.484142] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.484730] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b98132-71ca-4886-ae80-2179721f37ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.505552] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.505799] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27441d66-ebe6-45b0-a014-1e318a5f16ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.525607] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1156.525607] env[61995]: value = "task-795415" [ 1156.525607] env[61995]: _type = "Task" [ 1156.525607] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.533690] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795415, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.578696] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a73742da-3d5c-4b71-8e92-32e26d404f37] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.674472] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795412, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.779216] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795414, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.853314] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0aeae0-6248-4428-8a55-fe348bbffdd1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.860733] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d025396c-9988-4419-a7d5-ade884793a0c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.891553] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde5f770-6251-4475-9ab6-c96906fb6934 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.899133] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dfa824-d3a0-4035-b2e1-371bbe7b14ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.912940] env[61995]: DEBUG nova.compute.provider_tree [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1157.035728] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795415, 'name': ReconfigVM_Task, 'duration_secs': 0.307647} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.036108] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Reconfigured VM instance instance-0000006c to attach disk [datastore2] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1157.036747] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7df28eea-4981-4b9c-9f52-742457bd1bc4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.042982] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1157.042982] env[61995]: value = "task-795416" [ 1157.042982] env[61995]: _type = "Task" [ 1157.042982] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.051480] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795416, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.082452] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ac81c31d-d310-42c7-a6b4-166f3e82cf56] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.175650] env[61995]: DEBUG oslo_vmware.api [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795412, 'name': RemoveSnapshot_Task, 'duration_secs': 1.351845} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.175924] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1157.277405] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795414, 'name': CreateVM_Task, 'duration_secs': 0.652913} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.277727] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1157.278188] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.278431] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.278673] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1157.279399] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33343603-5798-40c2-bea3-333aefc647cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.283464] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1157.283464] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c4ccdf-0340-45a9-db42-684aa21e782e" [ 1157.283464] env[61995]: _type = "Task" [ 1157.283464] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.287233] env[61995]: INFO nova.compute.manager [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Swapping old allocation on dict_keys(['5c086f4d-bc91-4e49-9831-bed8df133c15']) held by migration 4c60573a-8da0-48d9-adde-dc3dacdd14d3 for instance [ 1157.293469] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c4ccdf-0340-45a9-db42-684aa21e782e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.308619] env[61995]: DEBUG nova.scheduler.client.report [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Overwriting current allocation {'allocations': {'5c086f4d-bc91-4e49-9831-bed8df133c15': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 148}}, 'project_id': '4c3e03b3825c43ca987d60b339fc0a52', 'user_id': '64e08719d59b456786fce984cc1508e4', 'consumer_generation': 1} on consumer b2aa6712-3284-4d97-ac98-ff8789fa8bca {{(pid=61995) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1157.379635] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.379831] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.380023] env[61995]: DEBUG nova.network.neutron [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1157.433573] env[61995]: ERROR nova.scheduler.client.report [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [req-45ea7f38-7fb9-4a13-bcf9-68f76f6802f6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-45ea7f38-7fb9-4a13-bcf9-68f76f6802f6"}]} [ 1157.448796] env[61995]: DEBUG nova.scheduler.client.report [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1157.461117] env[61995]: DEBUG nova.scheduler.client.report [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1157.461324] env[61995]: DEBUG nova.compute.provider_tree [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1157.470861] env[61995]: DEBUG nova.scheduler.client.report [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1157.486586] env[61995]: DEBUG nova.scheduler.client.report [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1157.545018] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef939042-c59c-4cab-9930-647c225dc5cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.557089] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d072de8-eea3-477a-9ef3-2a6fee921e79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.559932] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795416, 'name': Rename_Task, 'duration_secs': 0.142096} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.560209] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1157.560716] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06651b9f-4143-41dd-9a9f-a596a9ec4421 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.585308] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e2014437-fbcd-454a-893f-9ad7c7d461d0] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.588535] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0675e5ac-c6a2-440c-aa54-02bb691b51e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.590857] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1157.590857] env[61995]: value = "task-795417" [ 1157.590857] env[61995]: _type = "Task" [ 1157.590857] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.597620] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0feee702-798b-4bc1-8d99-1ed36479cc6b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.604194] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795417, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.613502] env[61995]: DEBUG nova.compute.provider_tree [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1157.681153] env[61995]: WARNING nova.compute.manager [None req-1ae72d7f-4176-4332-b244-7746aa0b6a12 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Image not found during snapshot: nova.exception.ImageNotFound: Image 5a9e44be-fb00-4d0c-b088-3150197a6c5a could not be found. [ 1157.793700] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c4ccdf-0340-45a9-db42-684aa21e782e, 'name': SearchDatastore_Task, 'duration_secs': 0.009822} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.793955] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.794217] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1157.794456] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.794605] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.794785] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1157.795049] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3981fc9-e0f1-4ab5-918d-86c88905c279 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.802753] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1157.802928] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1157.803610] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c400f34-ee15-4b83-a4dd-c0028e98a6fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.808746] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1157.808746] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529d79e1-99b0-c6a2-e5af-bd3815f0f69c" [ 1157.808746] env[61995]: _type = "Task" [ 1157.808746] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.816196] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529d79e1-99b0-c6a2-e5af-bd3815f0f69c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.092646] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: c3b2e000-cfa8-4b5b-8763-982c6b768621] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1158.104235] env[61995]: DEBUG oslo_vmware.api [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795417, 'name': PowerOnVM_Task, 'duration_secs': 0.427836} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.104563] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1158.104843] env[61995]: INFO nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Took 6.94 seconds to spawn the instance on the hypervisor. [ 1158.105058] env[61995]: DEBUG nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.105849] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f35a96a-ad27-45fc-bf8f-7eccf3a8cf0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.146062] env[61995]: DEBUG nova.scheduler.client.report [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 149 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1158.146339] env[61995]: DEBUG nova.compute.provider_tree [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 149 to 150 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1158.146754] env[61995]: DEBUG nova.compute.provider_tree [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1158.150443] env[61995]: DEBUG nova.network.neutron [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [{"id": "06334396-e435-480b-ab0c-0da719e15bf1", "address": "fa:16:3e:ba:41:b1", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06334396-e4", "ovs_interfaceid": "06334396-e435-480b-ab0c-0da719e15bf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.319892] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529d79e1-99b0-c6a2-e5af-bd3815f0f69c, 'name': SearchDatastore_Task, 'duration_secs': 0.008268} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.320852] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81cc537f-082d-44cc-914e-7c1d70aaa071 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.326679] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1158.326679] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5234c4d0-0081-ee59-9fe4-e490e77963ce" [ 1158.326679] env[61995]: _type = "Task" [ 1158.326679] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.334535] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5234c4d0-0081-ee59-9fe4-e490e77963ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.600053] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ff07d5b8-38b4-48d4-a6d0-632d37d3a91a] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1158.630987] env[61995]: INFO nova.compute.manager [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Took 13.44 seconds to build instance. [ 1158.652770] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.911s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.653362] env[61995]: DEBUG nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1158.656468] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-b2aa6712-3284-4d97-ac98-ff8789fa8bca" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.656468] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1158.656961] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd76e008-8cca-4a5e-bfc7-0a29bca8649e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.664074] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1158.664074] env[61995]: value = "task-795419" [ 1158.664074] env[61995]: _type = "Task" [ 1158.664074] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.672338] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795419, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.691063] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "1142205a-bc9e-4a94-b201-2fc203f8f913" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.691360] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.691583] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "1142205a-bc9e-4a94-b201-2fc203f8f913-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.691771] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.691954] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.695337] env[61995]: INFO nova.compute.manager [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Terminating instance [ 1158.697642] env[61995]: DEBUG nova.compute.manager [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1158.697861] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1158.698765] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c3c64a-8ebc-44dc-b3f9-13d3c54dfda1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.707519] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1158.707773] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25f3692d-f23d-4e79-afdc-c9c58b241d16 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.716919] env[61995]: DEBUG oslo_vmware.api [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1158.716919] env[61995]: value = "task-795420" [ 1158.716919] env[61995]: _type = "Task" [ 1158.716919] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.726745] env[61995]: DEBUG oslo_vmware.api [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.838706] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5234c4d0-0081-ee59-9fe4-e490e77963ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009961} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.839122] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.839544] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1158.839919] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f01ec1e-91d4-42ba-ad4d-0c963c8e7ff1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.847343] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1158.847343] env[61995]: value = "task-795421" [ 1158.847343] env[61995]: _type = "Task" [ 1158.847343] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.855379] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795421, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.103910] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 83369251-b00f-4595-bf98-28a3ec84f037] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.133220] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6aa44b6f-12ca-4913-8b0c-2af01c16f714 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.956s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.159249] env[61995]: DEBUG nova.compute.utils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1159.160824] env[61995]: DEBUG nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1159.161012] env[61995]: DEBUG nova.network.neutron [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1159.178821] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795419, 'name': PowerOffVM_Task, 'duration_secs': 0.195155} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.179191] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1159.179886] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1159.180499] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1159.180499] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1159.180499] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1159.180691] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1159.180760] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1159.180977] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1159.181161] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1159.181939] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1159.182388] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1159.182730] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1159.189792] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c6e1201-d6a2-4122-a74b-17c11d825a3a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.213716] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1159.213716] env[61995]: value = "task-795422" [ 1159.213716] env[61995]: _type = "Task" [ 1159.213716] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.229047] env[61995]: DEBUG oslo_vmware.api [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795420, 'name': PowerOffVM_Task, 'duration_secs': 0.217612} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.230671] env[61995]: DEBUG nova.policy [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e2f49fe94af402abe465e8415b16f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f3d88cb0eb54bc1a852c5f43f3806d9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1159.236530] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1159.236890] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1159.238798] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795422, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.239083] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77a024ce-3873-4466-9fda-0d461bc3a092 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.323653] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1159.323653] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1159.323653] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleting the datastore file [datastore2] 1142205a-bc9e-4a94-b201-2fc203f8f913 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1159.323653] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98df162b-f25a-4a5f-a08c-f7ca4edc35ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.332896] env[61995]: DEBUG oslo_vmware.api [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for the task: (returnval){ [ 1159.332896] env[61995]: value = "task-795424" [ 1159.332896] env[61995]: _type = "Task" [ 1159.332896] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.342929] env[61995]: DEBUG oslo_vmware.api [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.356619] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795421, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485869} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.356904] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1159.357163] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1159.357441] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8abaa53f-f8fb-4213-8347-5ac1b0345850 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.363274] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1159.363274] env[61995]: value = "task-795425" [ 1159.363274] env[61995]: _type = "Task" [ 1159.363274] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.371361] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.604015] env[61995]: DEBUG nova.network.neutron [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Successfully created port: 3ec98751-7ef3-4fa9-929a-f0b727483424 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1159.607474] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e8bde834-8f55-48d6-8623-1c6e43072c23] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.664267] env[61995]: DEBUG nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1159.724901] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795422, 'name': ReconfigVM_Task, 'duration_secs': 0.171866} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.725765] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e854450c-bcbd-4e97-9db7-3eb99522d63b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.748904] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1159.749962] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1159.750175] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1159.750853] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1159.750853] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1159.750853] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1159.751196] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1159.751196] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1159.751288] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1159.751404] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1159.751583] env[61995]: DEBUG nova.virt.hardware [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1159.752373] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc5b466c-39fc-438a-b1ec-eb9404564ddc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.758450] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1159.758450] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a69843-69eb-2407-2446-07cdacfbe610" [ 1159.758450] env[61995]: _type = "Task" [ 1159.758450] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.766405] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a69843-69eb-2407-2446-07cdacfbe610, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.843355] env[61995]: DEBUG oslo_vmware.api [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Task: {'id': task-795424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143568} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.843691] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1159.843890] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1159.844082] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1159.844266] env[61995]: INFO nova.compute.manager [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1159.844567] env[61995]: DEBUG oslo.service.loopingcall [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.844789] env[61995]: DEBUG nova.compute.manager [-] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1159.844882] env[61995]: DEBUG nova.network.neutron [-] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1159.873494] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068052} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.874106] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1159.874607] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8f9908-9743-44a6-b84e-3c1655ed979d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.899451] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.901948] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6ad02c3-5ca9-42ba-827f-677e93171a4b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.922679] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1159.922679] env[61995]: value = "task-795426" [ 1159.922679] env[61995]: _type = "Task" [ 1159.922679] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.933050] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.936980] env[61995]: DEBUG nova.compute.manager [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1159.937190] env[61995]: DEBUG nova.compute.manager [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing instance network info cache due to event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1159.937409] env[61995]: DEBUG oslo_concurrency.lockutils [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.937548] env[61995]: DEBUG oslo_concurrency.lockutils [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.937705] env[61995]: DEBUG nova.network.neutron [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1160.110614] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 410eac71-20cd-4a6d-9b78-e11fa72d74e7] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1160.268738] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a69843-69eb-2407-2446-07cdacfbe610, 'name': SearchDatastore_Task, 'duration_secs': 0.007209} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.274512] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1160.275681] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2011491-d33a-4b42-a2ff-490a5f842f02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.294506] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1160.294506] env[61995]: value = "task-795427" [ 1160.294506] env[61995]: _type = "Task" [ 1160.294506] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.306424] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795427, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.435081] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795426, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.619243] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a48d837d-1008-4248-bbe0-14321368e138] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1160.621035] env[61995]: DEBUG nova.network.neutron [-] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.676703] env[61995]: DEBUG nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1160.705515] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1160.705781] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1160.705939] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1160.706156] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1160.706377] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1160.706453] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1160.706663] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1160.706857] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1160.707083] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1160.707230] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1160.707420] env[61995]: DEBUG nova.virt.hardware [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1160.708384] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1410f8ec-aed7-4d41-9d7b-b2c9f0dcd6ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.718180] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba89aee3-7c29-4f5c-a731-f4976ab14d2e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.789485] env[61995]: DEBUG nova.network.neutron [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updated VIF entry in instance network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1160.789899] env[61995]: DEBUG nova.network.neutron [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.805439] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795427, 'name': ReconfigVM_Task, 'duration_secs': 0.401838} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.805687] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1160.806689] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7f2f39-178e-4285-913a-00cf05c30a3e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.830296] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1160.830606] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94fa3085-31ec-4056-af19-293ef586cfb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.849855] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1160.849855] env[61995]: value = "task-795428" [ 1160.849855] env[61995]: _type = "Task" [ 1160.849855] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.860394] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795428, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.934098] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795426, 'name': ReconfigVM_Task, 'duration_secs': 0.811422} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.934260] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b/1c2afe20-e28f-4553-a6e1-7f31ec0db64b.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.934882] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-593b9298-8db6-4f21-ac40-c8b6029591a3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.941937] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1160.941937] env[61995]: value = "task-795429" [ 1160.941937] env[61995]: _type = "Task" [ 1160.941937] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.951275] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795429, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.124758] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 432f4d05-ea4d-402c-8934-3aa1aaf93d48] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.127126] env[61995]: INFO nova.compute.manager [-] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Took 1.28 seconds to deallocate network for instance. [ 1161.183836] env[61995]: DEBUG nova.compute.manager [req-f6277290-8c9a-4d33-930a-508b864e0657 req-84f3bf18-5f9f-471f-94a7-8e93ff45f812 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Received event network-vif-plugged-3ec98751-7ef3-4fa9-929a-f0b727483424 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1161.184455] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6277290-8c9a-4d33-930a-508b864e0657 req-84f3bf18-5f9f-471f-94a7-8e93ff45f812 service nova] Acquiring lock "954d0ada-a870-415e-bd50-c066d27b7026-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.184455] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6277290-8c9a-4d33-930a-508b864e0657 req-84f3bf18-5f9f-471f-94a7-8e93ff45f812 service nova] Lock "954d0ada-a870-415e-bd50-c066d27b7026-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.184568] env[61995]: DEBUG oslo_concurrency.lockutils [req-f6277290-8c9a-4d33-930a-508b864e0657 req-84f3bf18-5f9f-471f-94a7-8e93ff45f812 service nova] Lock "954d0ada-a870-415e-bd50-c066d27b7026-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.184746] env[61995]: DEBUG nova.compute.manager [req-f6277290-8c9a-4d33-930a-508b864e0657 req-84f3bf18-5f9f-471f-94a7-8e93ff45f812 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] No waiting events found dispatching network-vif-plugged-3ec98751-7ef3-4fa9-929a-f0b727483424 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1161.184905] env[61995]: WARNING nova.compute.manager [req-f6277290-8c9a-4d33-930a-508b864e0657 req-84f3bf18-5f9f-471f-94a7-8e93ff45f812 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Received unexpected event network-vif-plugged-3ec98751-7ef3-4fa9-929a-f0b727483424 for instance with vm_state building and task_state spawning. [ 1161.292462] env[61995]: DEBUG oslo_concurrency.lockutils [req-5790de41-8dc8-4f25-8f9e-21ffbe58dcbf req-81d07f30-0ee7-40ac-966c-7c298cc9485b service nova] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.312884] env[61995]: DEBUG nova.network.neutron [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Successfully updated port: 3ec98751-7ef3-4fa9-929a-f0b727483424 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1161.359881] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795428, 'name': ReconfigVM_Task, 'duration_secs': 0.308252} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.360174] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Reconfigured VM instance instance-00000067 to attach disk [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca/b2aa6712-3284-4d97-ac98-ff8789fa8bca.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1161.361021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82299a68-c00f-455b-ac10-2036b9635321 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.380359] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d2471f-fe52-4b65-a949-a3de992bd28f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.398682] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e42964d-8120-4768-8057-a244d1e69005 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.416368] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc162480-cab5-4a28-a644-775cd25f82aa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.424321] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1161.424585] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69b27aa9-8a8a-4be0-b7cc-4700e2852c78 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.431286] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1161.431286] env[61995]: value = "task-795430" [ 1161.431286] env[61995]: _type = "Task" [ 1161.431286] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.438694] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795430, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.450678] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795429, 'name': Rename_Task, 'duration_secs': 0.14178} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.450977] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1161.451228] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f4d83df-ec32-4ebc-adb2-3b90191cf440 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.457238] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1161.457238] env[61995]: value = "task-795431" [ 1161.457238] env[61995]: _type = "Task" [ 1161.457238] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.466817] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795431, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.628656] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 06886222-5f7f-482e-b5ee-afd7326f2c70] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.636868] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.637147] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.637374] env[61995]: DEBUG nova.objects.instance [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lazy-loading 'resources' on Instance uuid 1142205a-bc9e-4a94-b201-2fc203f8f913 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.815067] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-954d0ada-a870-415e-bd50-c066d27b7026" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.815316] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-954d0ada-a870-415e-bd50-c066d27b7026" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.815411] env[61995]: DEBUG nova.network.neutron [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1161.946082] env[61995]: DEBUG oslo_vmware.api [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795430, 'name': PowerOnVM_Task, 'duration_secs': 0.386518} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.946082] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1161.969107] env[61995]: DEBUG oslo_vmware.api [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795431, 'name': PowerOnVM_Task, 'duration_secs': 0.431348} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.969570] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1161.969931] env[61995]: DEBUG nova.compute.manager [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1161.971818] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9b980c-6d1b-4188-9918-5cc3976a2ae4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.976181] env[61995]: DEBUG nova.compute.manager [req-be2f067c-3e6d-4ee9-b3cd-7d9509c7174e req-f161155f-1c73-4fac-89ee-3ce6251285d4 service nova] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Received event network-vif-deleted-f2a3f297-cb0b-4ba2-b472-65d810d6df37 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1162.134383] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 66c56f7b-1fc7-4aed-8afc-350817e1ca48] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.222026] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe16bab-2e1e-4273-a22e-77a6f413c16c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.228262] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f3e156-e5bb-4fd9-9239-2b6d01c01897 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.258882] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d2feb5-abc4-4bcc-aff9-df54823c1c70 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.266696] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08af9a3e-9822-42b4-96de-6988ffad68c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.281060] env[61995]: DEBUG nova.compute.provider_tree [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1162.357127] env[61995]: DEBUG nova.network.neutron [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1162.492014] env[61995]: DEBUG nova.network.neutron [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Updating instance_info_cache with network_info: [{"id": "3ec98751-7ef3-4fa9-929a-f0b727483424", "address": "fa:16:3e:93:32:e8", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ec98751-7e", "ovs_interfaceid": "3ec98751-7ef3-4fa9-929a-f0b727483424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.493220] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.636441] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 9e30eadd-5694-4fc7-8b54-2cf1d1571504] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.821215] env[61995]: DEBUG nova.scheduler.client.report [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 150 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1162.821505] env[61995]: DEBUG nova.compute.provider_tree [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 150 to 151 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1162.821690] env[61995]: DEBUG nova.compute.provider_tree [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1162.956967] env[61995]: INFO nova.compute.manager [None req-f7a82d85-5002-4125-9813-1f04907117e0 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance to original state: 'active' [ 1162.999265] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-954d0ada-a870-415e-bd50-c066d27b7026" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.999265] env[61995]: DEBUG nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Instance network_info: |[{"id": "3ec98751-7ef3-4fa9-929a-f0b727483424", "address": "fa:16:3e:93:32:e8", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ec98751-7e", "ovs_interfaceid": "3ec98751-7ef3-4fa9-929a-f0b727483424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1162.999265] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:32:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aef08290-001a-4ae8-aff0-1889e2211389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ec98751-7ef3-4fa9-929a-f0b727483424', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1163.007325] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating folder: Project (5f3d88cb0eb54bc1a852c5f43f3806d9). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1163.008572] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70060e86-2795-4b74-8a08-136cbc92c2ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.019364] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created folder: Project (5f3d88cb0eb54bc1a852c5f43f3806d9) in parent group-v185203. [ 1163.019557] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating folder: Instances. Parent ref: group-v185437. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1163.019796] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfdb14d9-aecf-455e-85f7-6e52535d475b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.030140] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created folder: Instances in parent group-v185437. [ 1163.030321] env[61995]: DEBUG oslo.service.loopingcall [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1163.030518] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1163.030947] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63b9c729-f96e-40c5-a501-41f2a220efb4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.049701] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1163.049701] env[61995]: value = "task-795434" [ 1163.049701] env[61995]: _type = "Task" [ 1163.049701] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.057242] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795434, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.139855] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 349667a3-6311-4c9d-bad2-ef4a5d1c52c8] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1163.212635] env[61995]: DEBUG nova.compute.manager [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Received event network-changed-3ec98751-7ef3-4fa9-929a-f0b727483424 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1163.212926] env[61995]: DEBUG nova.compute.manager [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Refreshing instance network info cache due to event network-changed-3ec98751-7ef3-4fa9-929a-f0b727483424. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1163.213888] env[61995]: DEBUG oslo_concurrency.lockutils [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] Acquiring lock "refresh_cache-954d0ada-a870-415e-bd50-c066d27b7026" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.213888] env[61995]: DEBUG oslo_concurrency.lockutils [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] Acquired lock "refresh_cache-954d0ada-a870-415e-bd50-c066d27b7026" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.214029] env[61995]: DEBUG nova.network.neutron [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Refreshing network info cache for port 3ec98751-7ef3-4fa9-929a-f0b727483424 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1163.327818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.691s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.330376] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.837s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.331080] env[61995]: DEBUG nova.objects.instance [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61995) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1163.353075] env[61995]: INFO nova.scheduler.client.report [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Deleted allocations for instance 1142205a-bc9e-4a94-b201-2fc203f8f913 [ 1163.437856] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.438186] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.438419] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.439018] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.439018] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.441051] env[61995]: INFO nova.compute.manager [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Terminating instance [ 1163.443419] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "refresh_cache-1c2afe20-e28f-4553-a6e1-7f31ec0db64b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.443584] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquired lock "refresh_cache-1c2afe20-e28f-4553-a6e1-7f31ec0db64b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.443817] env[61995]: DEBUG nova.network.neutron [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1163.560324] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795434, 'name': CreateVM_Task, 'duration_secs': 0.323981} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.560324] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1163.560707] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.560871] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.561231] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1163.561476] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5369224f-63a1-4ed3-94bd-9b7628167300 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.565853] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1163.565853] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]521f0f36-4c87-d8f6-b941-c9459d5e8adb" [ 1163.565853] env[61995]: _type = "Task" [ 1163.565853] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.573966] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521f0f36-4c87-d8f6-b941-c9459d5e8adb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.643359] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 9627f278-73d6-45d3-b60d-776e373eef73] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1163.861521] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e13c2792-3d99-4ea1-9ac5-749cb849dda9 tempest-ImagesTestJSON-80207239 tempest-ImagesTestJSON-80207239-project-member] Lock "1142205a-bc9e-4a94-b201-2fc203f8f913" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.170s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.919943] env[61995]: DEBUG nova.network.neutron [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Updated VIF entry in instance network info cache for port 3ec98751-7ef3-4fa9-929a-f0b727483424. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1163.920331] env[61995]: DEBUG nova.network.neutron [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Updating instance_info_cache with network_info: [{"id": "3ec98751-7ef3-4fa9-929a-f0b727483424", "address": "fa:16:3e:93:32:e8", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ec98751-7e", "ovs_interfaceid": "3ec98751-7ef3-4fa9-929a-f0b727483424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.962923] env[61995]: DEBUG nova.network.neutron [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1164.012466] env[61995]: DEBUG nova.network.neutron [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.075755] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]521f0f36-4c87-d8f6-b941-c9459d5e8adb, 'name': SearchDatastore_Task, 'duration_secs': 0.011198} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.076114] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.076440] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1164.076685] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.076837] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.077031] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1164.077302] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bee69d59-83b0-4264-a717-6c55f48bd1db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.085666] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1164.085850] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1164.086622] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fed80705-0a6a-45e4-8515-5f84fc3279ca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.091907] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1164.091907] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5234be9c-ca3a-1e46-70ea-92752cc48b26" [ 1164.091907] env[61995]: _type = "Task" [ 1164.091907] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.099793] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5234be9c-ca3a-1e46-70ea-92752cc48b26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.146620] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 3a2adf56-8071-476d-8cf8-5f6ece6a1ac4] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1164.341783] env[61995]: DEBUG oslo_concurrency.lockutils [None req-867974ef-8a52-416a-94d0-df1025097abb tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.423110] env[61995]: DEBUG oslo_concurrency.lockutils [req-508d79ec-70c6-417a-ae8d-5322df5dcbde req-a8a91ac6-9981-42b3-bc8d-8afa48045840 service nova] Releasing lock "refresh_cache-954d0ada-a870-415e-bd50-c066d27b7026" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.515085] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Releasing lock "refresh_cache-1c2afe20-e28f-4553-a6e1-7f31ec0db64b" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.515767] env[61995]: DEBUG nova.compute.manager [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1164.516171] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1164.517765] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9fdf9b-50af-45c3-8f93-d92f1ad3c144 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.526515] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1164.526773] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-076e9fb1-2600-45de-85fd-46666dea95d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.533095] env[61995]: DEBUG oslo_vmware.api [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1164.533095] env[61995]: value = "task-795435" [ 1164.533095] env[61995]: _type = "Task" [ 1164.533095] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.541564] env[61995]: DEBUG oslo_vmware.api [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.605171] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5234be9c-ca3a-1e46-70ea-92752cc48b26, 'name': SearchDatastore_Task, 'duration_secs': 0.010969} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.606190] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72a0c7b2-b93f-448d-ac7e-6d69403bd3cd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.612841] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1164.612841] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520a70bc-40a5-8379-b364-318747ede404" [ 1164.612841] env[61995]: _type = "Task" [ 1164.612841] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.624964] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520a70bc-40a5-8379-b364-318747ede404, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.650151] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 8ae65d63-16c7-4c67-a8c2-73968f63457d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.044512] env[61995]: DEBUG oslo_vmware.api [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795435, 'name': PowerOffVM_Task, 'duration_secs': 0.130582} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.044963] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1165.045272] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1165.045678] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e969d05-a13e-4ebe-a989-8f0c14b96d24 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.056312] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.057523] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.057523] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.057523] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.057523] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.061580] env[61995]: INFO nova.compute.manager [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Terminating instance [ 1165.063846] env[61995]: DEBUG nova.compute.manager [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1165.064165] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1165.065175] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1babf22f-da60-400f-8ccb-e655f7aa00e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.072975] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1165.073292] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85ff011f-ca97-434a-85f7-b14c6d68da83 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.076556] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1165.076703] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1165.076886] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Deleting the datastore file [datastore1] 1c2afe20-e28f-4553-a6e1-7f31ec0db64b {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1165.077414] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61c0b0f7-c6b2-44f3-a72a-51734ec52429 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.083253] env[61995]: DEBUG oslo_vmware.api [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1165.083253] env[61995]: value = "task-795437" [ 1165.083253] env[61995]: _type = "Task" [ 1165.083253] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.087021] env[61995]: DEBUG oslo_vmware.api [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for the task: (returnval){ [ 1165.087021] env[61995]: value = "task-795438" [ 1165.087021] env[61995]: _type = "Task" [ 1165.087021] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.093325] env[61995]: DEBUG oslo_vmware.api [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.099609] env[61995]: DEBUG oslo_vmware.api [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795438, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.123331] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520a70bc-40a5-8379-b364-318747ede404, 'name': SearchDatastore_Task, 'duration_secs': 0.012812} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.123599] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.123859] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 954d0ada-a870-415e-bd50-c066d27b7026/954d0ada-a870-415e-bd50-c066d27b7026.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1165.124133] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-115964c6-b1f4-4b1c-b51c-4878d72245eb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.132010] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1165.132010] env[61995]: value = "task-795439" [ 1165.132010] env[61995]: _type = "Task" [ 1165.132010] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.142027] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795439, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.153720] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ba89ee63-e293-47e1-90ab-7b8e72dd1b50] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.598691] env[61995]: DEBUG oslo_vmware.api [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795437, 'name': PowerOffVM_Task, 'duration_secs': 0.227936} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.598691] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1165.598691] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1165.598691] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e37ae7af-0ff1-4179-b813-3b0d315e6f51 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.604226] env[61995]: DEBUG oslo_vmware.api [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Task: {'id': task-795438, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131002} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.605484] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1165.605910] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1165.606235] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1165.606529] env[61995]: INFO nova.compute.manager [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 1165.606907] env[61995]: DEBUG oslo.service.loopingcall [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.607225] env[61995]: DEBUG nova.compute.manager [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1165.607421] env[61995]: DEBUG nova.network.neutron [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1165.641201] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795439, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.659529] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 847df13f-8451-49a1-9c6a-7c7c26952940] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.671021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1165.671021] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1165.671021] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleting the datastore file [datastore1] b2aa6712-3284-4d97-ac98-ff8789fa8bca {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1165.671021] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52401501-7160-4f98-a954-50f9e07f6630 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.678873] env[61995]: DEBUG oslo_vmware.api [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1165.678873] env[61995]: value = "task-795441" [ 1165.678873] env[61995]: _type = "Task" [ 1165.678873] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.688608] env[61995]: DEBUG oslo_vmware.api [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795441, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.733865] env[61995]: DEBUG nova.network.neutron [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1166.144143] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795439, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552745} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.144728] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 954d0ada-a870-415e-bd50-c066d27b7026/954d0ada-a870-415e-bd50-c066d27b7026.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1166.144728] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1166.144935] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0556c31-2726-4f5c-aa9f-fbdd29839fc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.151899] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1166.151899] env[61995]: value = "task-795442" [ 1166.151899] env[61995]: _type = "Task" [ 1166.151899] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.160335] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.162951] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 5cb4c1d9-e661-48b4-8600-1b823d63ddbb] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.189342] env[61995]: DEBUG oslo_vmware.api [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795441, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.403356} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.190214] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1166.190463] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1166.190603] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1166.190808] env[61995]: INFO nova.compute.manager [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1166.191093] env[61995]: DEBUG oslo.service.loopingcall [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1166.191745] env[61995]: DEBUG nova.compute.manager [-] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1166.191745] env[61995]: DEBUG nova.network.neutron [-] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1166.236322] env[61995]: DEBUG nova.network.neutron [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.664895] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077766} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.665290] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1166.665943] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9b4775-6254-4c2c-9512-c55a825d2622 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.670090] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 24b974c0-9567-46a3-96f6-fac1e602f46e] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.673646] env[61995]: DEBUG nova.compute.manager [req-bb2ebfec-c763-4a44-a8c1-6e4b32532607 req-32dcc874-7ddd-4ad0-82c7-cf625a87a678 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Received event network-vif-deleted-06334396-e435-480b-ab0c-0da719e15bf1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1166.673646] env[61995]: INFO nova.compute.manager [req-bb2ebfec-c763-4a44-a8c1-6e4b32532607 req-32dcc874-7ddd-4ad0-82c7-cf625a87a678 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Neutron deleted interface 06334396-e435-480b-ab0c-0da719e15bf1; detaching it from the instance and deleting it from the info cache [ 1166.673646] env[61995]: DEBUG nova.network.neutron [req-bb2ebfec-c763-4a44-a8c1-6e4b32532607 req-32dcc874-7ddd-4ad0-82c7-cf625a87a678 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.700721] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 954d0ada-a870-415e-bd50-c066d27b7026/954d0ada-a870-415e-bd50-c066d27b7026.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1166.701621] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebcf3a09-ac41-4f9e-b1b0-436e316c40c6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.725224] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1166.725224] env[61995]: value = "task-795443" [ 1166.725224] env[61995]: _type = "Task" [ 1166.725224] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.732886] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795443, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.740810] env[61995]: INFO nova.compute.manager [-] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Took 1.13 seconds to deallocate network for instance. [ 1167.143205] env[61995]: DEBUG nova.network.neutron [-] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.175472] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 453f46c7-5bad-4ca2-b228-f76e62fbd03e] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.179242] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54a46f97-e2df-47ac-8394-85fb19ca96a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.191022] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2c7c3b-0b4d-4321-a14d-ccd04d5628dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.214269] env[61995]: DEBUG nova.compute.manager [req-bb2ebfec-c763-4a44-a8c1-6e4b32532607 req-32dcc874-7ddd-4ad0-82c7-cf625a87a678 service nova] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Detach interface failed, port_id=06334396-e435-480b-ab0c-0da719e15bf1, reason: Instance b2aa6712-3284-4d97-ac98-ff8789fa8bca could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1167.234333] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795443, 'name': ReconfigVM_Task, 'duration_secs': 0.308037} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.234333] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 954d0ada-a870-415e-bd50-c066d27b7026/954d0ada-a870-415e-bd50-c066d27b7026.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1167.234595] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7752f98f-f5fb-45cc-a98d-1b008a1f38d5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.241476] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1167.241476] env[61995]: value = "task-795444" [ 1167.241476] env[61995]: _type = "Task" [ 1167.241476] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.247043] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.247189] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.247456] env[61995]: DEBUG nova.objects.instance [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lazy-loading 'resources' on Instance uuid 1c2afe20-e28f-4553-a6e1-7f31ec0db64b {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.253007] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795444, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.646639] env[61995]: INFO nova.compute.manager [-] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Took 1.45 seconds to deallocate network for instance. [ 1167.682478] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 88638f35-950d-42b4-92b2-ed0dced9148c] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.751218] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795444, 'name': Rename_Task, 'duration_secs': 0.232005} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.751504] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1167.751744] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd1851c9-5eda-4ff1-9e43-aab20f651379 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.758095] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1167.758095] env[61995]: value = "task-795445" [ 1167.758095] env[61995]: _type = "Task" [ 1167.758095] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.765480] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795445, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.821242] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719af602-b050-47f0-b486-be9a20eec3f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.828188] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6aba6e2-0d75-45b5-8525-d8729452859d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.858662] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cca7c72-aa4b-496e-9ae3-783ae9ad0d44 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.865788] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98dc9f0-3d5b-4ab1-8fdb-d0c201ba5f96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.883055] env[61995]: DEBUG nova.compute.provider_tree [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1168.156972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.185550] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: bbdcbc7d-d8a9-4ae2-9451-5edbb59b4a31] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.268761] env[61995]: DEBUG oslo_vmware.api [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795445, 'name': PowerOnVM_Task, 'duration_secs': 0.500099} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.269041] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1168.269306] env[61995]: INFO nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Took 7.59 seconds to spawn the instance on the hypervisor. [ 1168.269481] env[61995]: DEBUG nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1168.270296] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3992146f-c3a4-40ad-a65c-b20bc4c8bbb3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.424380] env[61995]: DEBUG nova.scheduler.client.report [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 151 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1168.424687] env[61995]: DEBUG nova.compute.provider_tree [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 151 to 152 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1168.424880] env[61995]: DEBUG nova.compute.provider_tree [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1168.690878] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 057ba9d6-4ff2-4da1-be7d-ecfadf738d74] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.792773] env[61995]: INFO nova.compute.manager [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Took 18.08 seconds to build instance. [ 1168.931157] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.935361] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.777s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.935361] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.954255] env[61995]: INFO nova.scheduler.client.report [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted allocations for instance b2aa6712-3284-4d97-ac98-ff8789fa8bca [ 1168.955892] env[61995]: INFO nova.scheduler.client.report [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Deleted allocations for instance 1c2afe20-e28f-4553-a6e1-7f31ec0db64b [ 1169.193621] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7d987006-afac-4973-ae69-f15b4752469b] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.295893] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4233ec9f-e4fc-450e-9882-eff2b836d7be tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "954d0ada-a870-415e-bd50-c066d27b7026" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.597s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.466712] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bf711276-4571-4be7-94d3-1f1a9aeed245 tempest-ServerShowV254Test-65464125 tempest-ServerShowV254Test-65464125-project-member] Lock "1c2afe20-e28f-4553-a6e1-7f31ec0db64b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.028s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.467777] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ad1c58f3-fc18-4777-a2ee-edddac17380c tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "b2aa6712-3284-4d97-ac98-ff8789fa8bca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.411s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.698239] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 5a6bd8a7-2633-42d2-b853-4e3a5690c77e] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.904912] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "954d0ada-a870-415e-bd50-c066d27b7026" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.905255] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "954d0ada-a870-415e-bd50-c066d27b7026" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.905506] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "954d0ada-a870-415e-bd50-c066d27b7026-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.905713] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "954d0ada-a870-415e-bd50-c066d27b7026-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.905896] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "954d0ada-a870-415e-bd50-c066d27b7026-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.908254] env[61995]: INFO nova.compute.manager [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Terminating instance [ 1169.912228] env[61995]: DEBUG nova.compute.manager [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1169.912228] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1169.913175] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d942fef-1459-4267-b8ef-f0d689778381 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.922794] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1169.922794] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21cfe5b0-2a31-4f80-bf34-857f3f5c772b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.929249] env[61995]: DEBUG oslo_vmware.api [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1169.929249] env[61995]: value = "task-795446" [ 1169.929249] env[61995]: _type = "Task" [ 1169.929249] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.937569] env[61995]: DEBUG oslo_vmware.api [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795446, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.962473] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.962729] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.195437] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.195665] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.199640] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a8b7bcd6-4753-4e1f-8b0b-777c882b99f5] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1170.439219] env[61995]: DEBUG oslo_vmware.api [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795446, 'name': PowerOffVM_Task, 'duration_secs': 0.177332} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.439672] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1170.439959] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1170.440344] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b30e43b9-fac8-4c92-8406-e09397494ef0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.465195] env[61995]: DEBUG nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1170.505348] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1170.505583] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1170.505771] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleting the datastore file [datastore2] 954d0ada-a870-415e-bd50-c066d27b7026 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.506056] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f3da594-658a-4d01-826b-ed2b27de21fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.513459] env[61995]: DEBUG oslo_vmware.api [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1170.513459] env[61995]: value = "task-795448" [ 1170.513459] env[61995]: _type = "Task" [ 1170.513459] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.521709] env[61995]: DEBUG oslo_vmware.api [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795448, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.698249] env[61995]: DEBUG nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1170.705932] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e835ac53-29af-4bd0-b186-5c6270ccf760] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1170.986057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.986333] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.987877] env[61995]: INFO nova.compute.claims [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1171.023343] env[61995]: DEBUG oslo_vmware.api [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795448, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125681} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.024224] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.024422] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1171.024601] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1171.024775] env[61995]: INFO nova.compute.manager [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1171.025043] env[61995]: DEBUG oslo.service.loopingcall [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.025259] env[61995]: DEBUG nova.compute.manager [-] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1171.025353] env[61995]: DEBUG nova.network.neutron [-] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1171.209418] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 4ca0dcd7-a2b8-4de6-835b-07c36ab95a99] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.217793] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.281054] env[61995]: DEBUG nova.compute.manager [req-333a3275-22de-41c4-b171-05ddaacfeed7 req-b65b31ba-c5a5-406a-95bc-c66695f29916 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Received event network-vif-deleted-3ec98751-7ef3-4fa9-929a-f0b727483424 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1171.281318] env[61995]: INFO nova.compute.manager [req-333a3275-22de-41c4-b171-05ddaacfeed7 req-b65b31ba-c5a5-406a-95bc-c66695f29916 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Neutron deleted interface 3ec98751-7ef3-4fa9-929a-f0b727483424; detaching it from the instance and deleting it from the info cache [ 1171.281500] env[61995]: DEBUG nova.network.neutron [req-333a3275-22de-41c4-b171-05ddaacfeed7 req-b65b31ba-c5a5-406a-95bc-c66695f29916 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.712634] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a647b9e7-dd9c-48a3-a0a2-c06aa5611fc5] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.759794] env[61995]: DEBUG nova.network.neutron [-] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.783954] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0a82c01d-96c5-4692-bbe2-b7d9874e93d6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.795256] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a9f037-e7f4-4885-a0c5-a1a127a405ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.818224] env[61995]: DEBUG nova.compute.manager [req-333a3275-22de-41c4-b171-05ddaacfeed7 req-b65b31ba-c5a5-406a-95bc-c66695f29916 service nova] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Detach interface failed, port_id=3ec98751-7ef3-4fa9-929a-f0b727483424, reason: Instance 954d0ada-a870-415e-bd50-c066d27b7026 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1172.057771] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150bb7fd-0eba-4cb6-98d5-04c8e9469eed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.065516] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e62115-b908-44a1-a5de-d9b97e4a2e74 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.094093] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c21429-07b6-4111-8c01-15dc6682b979 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.100692] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fc26e0-9b8d-4ff3-a43b-72f470c25565 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.112885] env[61995]: DEBUG nova.compute.provider_tree [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.216488] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: fdeceddf-c6f9-4a98-afd0-defa1284eeff] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.262774] env[61995]: INFO nova.compute.manager [-] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Took 1.24 seconds to deallocate network for instance. [ 1172.617100] env[61995]: DEBUG nova.scheduler.client.report [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1172.719271] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 4b49e7b1-7fd0-4e59-af79-5c0898967c35] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.768744] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.122519] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.136s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.123057] env[61995]: DEBUG nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1173.125871] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.908s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.127396] env[61995]: INFO nova.compute.claims [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1173.222264] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ed952baf-f1ef-4ec0-8a15-928287190456] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1173.632021] env[61995]: DEBUG nova.compute.utils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1173.635437] env[61995]: DEBUG nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1173.635606] env[61995]: DEBUG nova.network.neutron [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1173.684577] env[61995]: DEBUG nova.policy [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8feed88faa214a6fa5c51f6f0cca4ea7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49e78af0ad9340258211bf92e447021c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1173.725366] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: cfcf1a5a-97db-4095-b6a6-044242ee8a05] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1173.977660] env[61995]: DEBUG nova.network.neutron [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Successfully created port: 47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1174.139591] env[61995]: DEBUG nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1174.219114] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac066013-d362-41f1-9534-9d2b11e227f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.226940] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79dff51-b763-4995-9bd5-4afdfcde1863 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.230240] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 229d966c-b70d-4a51-a176-2e88488d5c4b] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1174.262147] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7275915d-a77c-477d-932c-c5f53a437e1e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.270192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1cdda8-3f6b-4cfb-9462-53b8a277b275 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.283976] env[61995]: DEBUG nova.compute.provider_tree [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.733621] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 5480efdf-fd46-4a26-b6a5-b6c3dd4c4400] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1174.786903] env[61995]: DEBUG nova.scheduler.client.report [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1175.152780] env[61995]: DEBUG nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1175.177285] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1175.177542] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1175.177701] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1175.177886] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1175.178051] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1175.178264] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1175.178489] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1175.178656] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1175.178827] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1175.178995] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1175.179188] env[61995]: DEBUG nova.virt.hardware [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1175.180046] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088d5759-5356-4670-afd9-7cc55d1d0868 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.187764] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3e3ddf-031f-4479-b95d-4d21ee4c400c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.236701] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 2b40609a-0826-4a16-af28-c8926af6b646] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1175.292119] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.292761] env[61995]: DEBUG nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1175.295871] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.527s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.296152] env[61995]: DEBUG nova.objects.instance [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'resources' on Instance uuid 954d0ada-a870-415e-bd50-c066d27b7026 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.363237] env[61995]: DEBUG nova.compute.manager [req-1140be1f-01b4-4563-a127-80bb4705bfe5 req-d84112dc-1ada-47f4-95c5-6018aa7207d8 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-vif-plugged-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1175.363237] env[61995]: DEBUG oslo_concurrency.lockutils [req-1140be1f-01b4-4563-a127-80bb4705bfe5 req-d84112dc-1ada-47f4-95c5-6018aa7207d8 service nova] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.363237] env[61995]: DEBUG oslo_concurrency.lockutils [req-1140be1f-01b4-4563-a127-80bb4705bfe5 req-d84112dc-1ada-47f4-95c5-6018aa7207d8 service nova] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.363237] env[61995]: DEBUG oslo_concurrency.lockutils [req-1140be1f-01b4-4563-a127-80bb4705bfe5 req-d84112dc-1ada-47f4-95c5-6018aa7207d8 service nova] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.363237] env[61995]: DEBUG nova.compute.manager [req-1140be1f-01b4-4563-a127-80bb4705bfe5 req-d84112dc-1ada-47f4-95c5-6018aa7207d8 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] No waiting events found dispatching network-vif-plugged-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1175.363237] env[61995]: WARNING nova.compute.manager [req-1140be1f-01b4-4563-a127-80bb4705bfe5 req-d84112dc-1ada-47f4-95c5-6018aa7207d8 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received unexpected event network-vif-plugged-47ddce51-cfec-49c4-ba3b-b34c86140652 for instance with vm_state building and task_state spawning. [ 1175.452447] env[61995]: DEBUG nova.network.neutron [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Successfully updated port: 47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1175.740137] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d3f629c5-96f4-4208-be0d-cbf9810ecc6b] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1175.799304] env[61995]: DEBUG nova.compute.utils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1175.803254] env[61995]: DEBUG nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1175.803420] env[61995]: DEBUG nova.network.neutron [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1175.841984] env[61995]: DEBUG nova.policy [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e08719d59b456786fce984cc1508e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c3e03b3825c43ca987d60b339fc0a52', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1175.865704] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de68206-f0b3-49b5-823e-9abab31b96dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.873321] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2a0a3f-db40-4ba5-a4df-ab3b3337858c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.902664] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d6c683-7eac-4a63-a45a-ce61da2dfe6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.911751] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade30773-682b-46de-9134-af0128b62e78 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.924238] env[61995]: DEBUG nova.compute.provider_tree [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.954622] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.954778] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.954926] env[61995]: DEBUG nova.network.neutron [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1176.167071] env[61995]: DEBUG nova.network.neutron [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Successfully created port: 2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1176.243634] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 966a72d8-a51a-44f9-8184-8108f8cc3ce6] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1176.303617] env[61995]: DEBUG nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1176.426979] env[61995]: DEBUG nova.scheduler.client.report [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.490046] env[61995]: DEBUG nova.network.neutron [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1176.607808] env[61995]: DEBUG nova.network.neutron [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47ddce51-cf", "ovs_interfaceid": "47ddce51-cfec-49c4-ba3b-b34c86140652", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.746280] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 451e226c-5fb7-4042-a706-3e58b0606546] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1176.931706] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.636s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.955713] env[61995]: INFO nova.scheduler.client.report [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted allocations for instance 954d0ada-a870-415e-bd50-c066d27b7026 [ 1177.110861] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.111214] env[61995]: DEBUG nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Instance network_info: |[{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47ddce51-cf", "ovs_interfaceid": "47ddce51-cfec-49c4-ba3b-b34c86140652", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1177.111637] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:74:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '28d04eee-6dbb-491a-a999-b659c799679d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47ddce51-cfec-49c4-ba3b-b34c86140652', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1177.119723] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating folder: Project (49e78af0ad9340258211bf92e447021c). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1177.119995] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7dc87cd4-8116-45d9-8fbf-ab3a90887014 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.133161] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created folder: Project (49e78af0ad9340258211bf92e447021c) in parent group-v185203. [ 1177.133387] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating folder: Instances. Parent ref: group-v185440. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1177.133690] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2da3f3e3-6c85-498f-a4e0-f1920a18bcab {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.144198] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created folder: Instances in parent group-v185440. [ 1177.144240] env[61995]: DEBUG oslo.service.loopingcall [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1177.144397] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1177.144587] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8290d4a4-be8a-4777-8fa7-634771272e01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.162836] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1177.162836] env[61995]: value = "task-795451" [ 1177.162836] env[61995]: _type = "Task" [ 1177.162836] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.170056] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795451, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.249896] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: af7163e8-4e77-4336-80e9-0623e29ed8c0] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1177.312365] env[61995]: DEBUG nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1177.333738] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1177.333964] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1177.334670] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1177.335030] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1177.335224] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1177.335383] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1177.335605] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1177.335769] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1177.335957] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1177.336142] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1177.336320] env[61995]: DEBUG nova.virt.hardware [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1177.337184] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec32df72-076d-400b-93d5-0f6ccfda81b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.346166] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e382e7-4ec1-4b09-b30f-459f9318534d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.391378] env[61995]: DEBUG nova.compute.manager [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1177.392137] env[61995]: DEBUG nova.compute.manager [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing instance network info cache due to event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1177.392137] env[61995]: DEBUG oslo_concurrency.lockutils [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] Acquiring lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.392137] env[61995]: DEBUG oslo_concurrency.lockutils [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] Acquired lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.392270] env[61995]: DEBUG nova.network.neutron [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1177.463626] env[61995]: DEBUG oslo_concurrency.lockutils [None req-50c6766f-86f3-49aa-b1b0-a5ef3b4ac1ec tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "954d0ada-a870-415e-bd50-c066d27b7026" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.558s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.655388] env[61995]: DEBUG nova.network.neutron [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Successfully updated port: 2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1177.673421] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795451, 'name': CreateVM_Task, 'duration_secs': 0.302861} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.673640] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1177.674353] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.674535] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.674857] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1177.675117] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb56aa2-3467-41ef-bd3a-d9ea6e078eda {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.679791] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1177.679791] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]525cff94-266e-b484-804e-143c4f0e0754" [ 1177.679791] env[61995]: _type = "Task" [ 1177.679791] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.687586] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525cff94-266e-b484-804e-143c4f0e0754, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.755049] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: bb31ef64-478f-4f39-9ee2-fcfabe8235ec] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.078203] env[61995]: DEBUG nova.network.neutron [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updated VIF entry in instance network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1178.078591] env[61995]: DEBUG nova.network.neutron [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47ddce51-cf", "ovs_interfaceid": "47ddce51-cfec-49c4-ba3b-b34c86140652", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.158352] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.158512] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.158661] env[61995]: DEBUG nova.network.neutron [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1178.164087] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "a3e4c652-49dc-48a1-affc-090e2b92736a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.164419] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.190761] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]525cff94-266e-b484-804e-143c4f0e0754, 'name': SearchDatastore_Task, 'duration_secs': 0.009112} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.191081] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.191329] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.191572] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.191738] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.191945] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.192212] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04858e72-5bc1-4341-a2ac-bb459bb4414f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.200893] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.201164] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1178.202040] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7735db5-9855-4438-ba32-2169253c0911 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.208062] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1178.208062] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52fc1805-7c3e-f66a-31b8-3232e5ee3851" [ 1178.208062] env[61995]: _type = "Task" [ 1178.208062] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.217474] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fc1805-7c3e-f66a-31b8-3232e5ee3851, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.258289] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7524c912-4d73-462a-949f-3b4c85874094] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.581274] env[61995]: DEBUG oslo_concurrency.lockutils [req-db129fdf-54bd-400c-9d8e-b6bfcca18895 req-9ef40726-d631-4eec-b9cf-7269bb79a825 service nova] Releasing lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.666993] env[61995]: DEBUG nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1178.696157] env[61995]: DEBUG nova.network.neutron [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1178.719692] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52fc1805-7c3e-f66a-31b8-3232e5ee3851, 'name': SearchDatastore_Task, 'duration_secs': 0.008415} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.720465] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebc73faa-d39d-41ee-9bb7-44db8c8dfda5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.725691] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1178.725691] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d3933d-579d-4eac-f674-60df81fd7ed4" [ 1178.725691] env[61995]: _type = "Task" [ 1178.725691] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.732698] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d3933d-579d-4eac-f674-60df81fd7ed4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.761605] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 0348098c-fd4c-41f8-a246-b9ba87a75674] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.822856] env[61995]: DEBUG nova.network.neutron [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updating instance_info_cache with network_info: [{"id": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "address": "fa:16:3e:3c:5e:4b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a2dabdb-39", "ovs_interfaceid": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.189269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.189572] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.191027] env[61995]: INFO nova.compute.claims [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1179.237139] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d3933d-579d-4eac-f674-60df81fd7ed4, 'name': SearchDatastore_Task, 'duration_secs': 0.009166} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.237425] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.237695] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1179.237956] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bba7dc10-566c-4ba2-9d85-e201ff247ac6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.245308] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1179.245308] env[61995]: value = "task-795452" [ 1179.245308] env[61995]: _type = "Task" [ 1179.245308] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.252726] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.264812] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 0c45e1fc-7b9a-4873-a381-ee2850ad5d4d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1179.325022] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.325098] env[61995]: DEBUG nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Instance network_info: |[{"id": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "address": "fa:16:3e:3c:5e:4b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a2dabdb-39", "ovs_interfaceid": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1179.325500] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:5e:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ada35c98-01a9-4352-98e4-1d20ba31f928', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a2dabdb-39a2-4193-accd-c3fb7fb129b9', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1179.333014] env[61995]: DEBUG oslo.service.loopingcall [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1179.333234] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1179.333460] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c0c0d97-e392-4501-85c7-9c5cd49cb90b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.355386] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1179.355386] env[61995]: value = "task-795453" [ 1179.355386] env[61995]: _type = "Task" [ 1179.355386] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.363321] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795453, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.453949] env[61995]: DEBUG nova.compute.manager [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Received event network-vif-plugged-2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1179.454355] env[61995]: DEBUG oslo_concurrency.lockutils [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] Acquiring lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.454438] env[61995]: DEBUG oslo_concurrency.lockutils [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.454695] env[61995]: DEBUG oslo_concurrency.lockutils [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.454759] env[61995]: DEBUG nova.compute.manager [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] No waiting events found dispatching network-vif-plugged-2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1179.454934] env[61995]: WARNING nova.compute.manager [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Received unexpected event network-vif-plugged-2a2dabdb-39a2-4193-accd-c3fb7fb129b9 for instance with vm_state building and task_state spawning. [ 1179.455277] env[61995]: DEBUG nova.compute.manager [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Received event network-changed-2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1179.455496] env[61995]: DEBUG nova.compute.manager [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Refreshing instance network info cache due to event network-changed-2a2dabdb-39a2-4193-accd-c3fb7fb129b9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1179.455691] env[61995]: DEBUG oslo_concurrency.lockutils [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] Acquiring lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.455831] env[61995]: DEBUG oslo_concurrency.lockutils [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] Acquired lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.456226] env[61995]: DEBUG nova.network.neutron [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Refreshing network info cache for port 2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1179.756071] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.405102} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.756071] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1179.756071] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1179.756334] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b33f5390-e486-49e1-bd6a-7856dbbe213e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.762616] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1179.762616] env[61995]: value = "task-795454" [ 1179.762616] env[61995]: _type = "Task" [ 1179.762616] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.767105] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: b40b1866-ced3-40ef-9ab7-30d58b75f288] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1179.771644] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795454, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.865010] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795453, 'name': CreateVM_Task, 'duration_secs': 0.469756} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.865229] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1179.865869] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.866061] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.866386] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1179.866702] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-571587d7-fd54-4463-b01c-f8de62b143e3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.870953] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1179.870953] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52784ff9-ee8e-250a-cdab-fad90f3f2815" [ 1179.870953] env[61995]: _type = "Task" [ 1179.870953] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.877997] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52784ff9-ee8e-250a-cdab-fad90f3f2815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.136264] env[61995]: DEBUG nova.network.neutron [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updated VIF entry in instance network info cache for port 2a2dabdb-39a2-4193-accd-c3fb7fb129b9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1180.136630] env[61995]: DEBUG nova.network.neutron [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updating instance_info_cache with network_info: [{"id": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "address": "fa:16:3e:3c:5e:4b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a2dabdb-39", "ovs_interfaceid": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.256737] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe20d9c7-0e59-47fa-a20b-81567af57a82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.266861] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f18e58-34a8-439e-8fae-007bd6c8fca7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.274363] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: dc7077ac-d3fd-4e84-867a-d86328b32f5d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1180.276155] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795454, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068527} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.301149] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1180.301969] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd99f4d-5739-488e-be2a-e5ae1f40867a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.304739] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5b1959-3e5a-49df-a59d-e45f1a92954a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.320563] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8078e4-5495-42cb-b502-beb4163b85be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.332341] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.332772] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-811f67a9-94a9-46b9-8ec6-1e8f8cd7a83e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.355973] env[61995]: DEBUG nova.compute.provider_tree [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.357886] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1180.357886] env[61995]: value = "task-795455" [ 1180.357886] env[61995]: _type = "Task" [ 1180.357886] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.365711] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795455, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.380235] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52784ff9-ee8e-250a-cdab-fad90f3f2815, 'name': SearchDatastore_Task, 'duration_secs': 0.007723} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.380511] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.380740] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1180.380971] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.381134] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.381319] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1180.381557] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1eaff827-6b7a-433f-aacc-c462612dc1a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.389472] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1180.389651] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1180.390305] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de8b1194-bffd-4d72-ac71-d61f745f61ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.394953] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1180.394953] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5246f616-76dc-09e0-a040-8e4cf8ea94ea" [ 1180.394953] env[61995]: _type = "Task" [ 1180.394953] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.401990] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5246f616-76dc-09e0-a040-8e4cf8ea94ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.639275] env[61995]: DEBUG oslo_concurrency.lockutils [req-d922039e-9a69-423c-9ff8-26ce7337f48d req-6e4d58c4-6b94-4778-b72b-09a446930de3 service nova] Releasing lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.777368] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 0e152049-f9ca-4ac1-b12d-ae1ad4623ff5] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1180.859857] env[61995]: DEBUG nova.scheduler.client.report [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1180.871655] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795455, 'name': ReconfigVM_Task, 'duration_secs': 0.275346} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.871924] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.872520] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3bf20ee4-2e34-4be7-ad23-9c9b2c5638b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.879698] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1180.879698] env[61995]: value = "task-795456" [ 1180.879698] env[61995]: _type = "Task" [ 1180.879698] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.893107] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795456, 'name': Rename_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.904605] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5246f616-76dc-09e0-a040-8e4cf8ea94ea, 'name': SearchDatastore_Task, 'duration_secs': 0.007524} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.905490] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a337037c-2080-43f8-bfd6-c05b574d1c31 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.910425] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1180.910425] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522ccdb5-e401-0bc8-f069-e75685b7fb5f" [ 1180.910425] env[61995]: _type = "Task" [ 1180.910425] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.918210] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522ccdb5-e401-0bc8-f069-e75685b7fb5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.280906] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: eef1cf32-e71e-4696-8170-1023e3a5518d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1181.367684] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.368251] env[61995]: DEBUG nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1181.388908] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795456, 'name': Rename_Task, 'duration_secs': 0.139709} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.388908] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1181.389139] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6eac9626-41e1-4c48-acbc-86ba5802331d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.396984] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1181.396984] env[61995]: value = "task-795457" [ 1181.396984] env[61995]: _type = "Task" [ 1181.396984] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.404487] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.419301] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522ccdb5-e401-0bc8-f069-e75685b7fb5f, 'name': SearchDatastore_Task, 'duration_secs': 0.009873} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.419616] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.419989] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] c99b0295-bf1f-4a25-af31-cd0e19b6b089/c99b0295-bf1f-4a25-af31-cd0e19b6b089.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1181.420270] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d99aff9f-8d88-4866-866e-946fd7c751ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.426639] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1181.426639] env[61995]: value = "task-795458" [ 1181.426639] env[61995]: _type = "Task" [ 1181.426639] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.435911] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.784528] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: f9c1424f-6096-481c-aa91-acc25bfd8f6e] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1181.873236] env[61995]: DEBUG nova.compute.utils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1181.875058] env[61995]: DEBUG nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1181.875058] env[61995]: DEBUG nova.network.neutron [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1181.906323] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795457, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.912422] env[61995]: DEBUG nova.policy [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e2f49fe94af402abe465e8415b16f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f3d88cb0eb54bc1a852c5f43f3806d9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1181.935689] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448722} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.935931] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] c99b0295-bf1f-4a25-af31-cd0e19b6b089/c99b0295-bf1f-4a25-af31-cd0e19b6b089.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1181.936168] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1181.936419] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1cc0f6b8-78d2-44ed-a90c-2f5749641d36 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.941984] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1181.941984] env[61995]: value = "task-795459" [ 1181.941984] env[61995]: _type = "Task" [ 1181.941984] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.949331] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795459, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.170389] env[61995]: DEBUG nova.network.neutron [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Successfully created port: 4f406016-d02c-46a1-915f-44e02ed7ad90 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1182.288442] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 4edbf65d-eaca-47e1-b6bf-03ccb908b52e] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1182.378090] env[61995]: DEBUG nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1182.407323] env[61995]: DEBUG oslo_vmware.api [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795457, 'name': PowerOnVM_Task, 'duration_secs': 0.698395} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.407696] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1182.408053] env[61995]: INFO nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1182.408336] env[61995]: DEBUG nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1182.409225] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b10b06-3896-4bc6-8d14-d19058c7f2ad {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.453350] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795459, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059282} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.454057] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1182.454595] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d040206b-8936-48ae-a4d5-ef57f07cb6e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.476159] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] c99b0295-bf1f-4a25-af31-cd0e19b6b089/c99b0295-bf1f-4a25-af31-cd0e19b6b089.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1182.476390] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-438870af-7d7e-4d1d-8ac1-62becda63d98 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.495293] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1182.495293] env[61995]: value = "task-795460" [ 1182.495293] env[61995]: _type = "Task" [ 1182.495293] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.502591] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795460, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.791583] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: fcdc7c33-8c46-4f4d-bb39-e6f943256cb0] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1182.925621] env[61995]: INFO nova.compute.manager [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Took 11.96 seconds to build instance. [ 1183.006180] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795460, 'name': ReconfigVM_Task, 'duration_secs': 0.273136} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.006483] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Reconfigured VM instance instance-0000006f to attach disk [datastore2] c99b0295-bf1f-4a25-af31-cd0e19b6b089/c99b0295-bf1f-4a25-af31-cd0e19b6b089.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1183.007133] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fdef5708-c4e7-45b5-9217-a7f54dd1b888 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.013711] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1183.013711] env[61995]: value = "task-795461" [ 1183.013711] env[61995]: _type = "Task" [ 1183.013711] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.021137] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795461, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.295217] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: c4f744f9-36f3-4913-9ced-f0db93d3b4df] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1183.300791] env[61995]: DEBUG nova.compute.manager [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1183.300791] env[61995]: DEBUG nova.compute.manager [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing instance network info cache due to event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1183.300791] env[61995]: DEBUG oslo_concurrency.lockutils [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] Acquiring lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.300791] env[61995]: DEBUG oslo_concurrency.lockutils [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] Acquired lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.300791] env[61995]: DEBUG nova.network.neutron [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1183.388088] env[61995]: DEBUG nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1183.427982] env[61995]: DEBUG oslo_concurrency.lockutils [None req-bfe40977-80d8-403a-a5af-9a5220ad59f8 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.465s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.482009] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1183.482278] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1183.482442] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1183.482626] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1183.482775] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1183.482923] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1183.483154] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1183.483321] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1183.483493] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1183.483661] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1183.483838] env[61995]: DEBUG nova.virt.hardware [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1183.484714] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92806bcf-7fe4-416b-b26e-5d25f99c6498 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.492576] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fb9abe-0d2a-40ff-aa6c-4ed095739118 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.522764] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795461, 'name': Rename_Task, 'duration_secs': 0.147895} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.523039] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1183.523277] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c20aef82-8076-418f-829b-d7c0bf56b44f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.530064] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1183.530064] env[61995]: value = "task-795462" [ 1183.530064] env[61995]: _type = "Task" [ 1183.530064] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.538585] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.651249] env[61995]: DEBUG nova.network.neutron [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Successfully updated port: 4f406016-d02c-46a1-915f-44e02ed7ad90 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1183.799521] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: b1bd98d9-bd0f-4abd-a188-e5267ada4852] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1184.024481] env[61995]: DEBUG nova.network.neutron [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updated VIF entry in instance network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1184.024858] env[61995]: DEBUG nova.network.neutron [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47ddce51-cf", "ovs_interfaceid": "47ddce51-cfec-49c4-ba3b-b34c86140652", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.039350] env[61995]: DEBUG oslo_vmware.api [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795462, 'name': PowerOnVM_Task, 'duration_secs': 0.445674} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.040227] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1184.040461] env[61995]: INFO nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Took 6.73 seconds to spawn the instance on the hypervisor. [ 1184.040649] env[61995]: DEBUG nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.041431] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cb8bf6-a7d1-4640-8e74-7c9f9ed2b53b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.153666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-a3e4c652-49dc-48a1-affc-090e2b92736a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.153825] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-a3e4c652-49dc-48a1-affc-090e2b92736a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.153975] env[61995]: DEBUG nova.network.neutron [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1184.302575] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: d9c53806-68c0-4872-a262-36bc05573674] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1184.528300] env[61995]: DEBUG oslo_concurrency.lockutils [req-45c35448-55b3-4808-b9ba-2daca044359a req-100d80f1-35df-423e-a915-623df9655fcb service nova] Releasing lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.560186] env[61995]: INFO nova.compute.manager [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Took 13.36 seconds to build instance. [ 1184.685574] env[61995]: DEBUG nova.network.neutron [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1184.805912] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: e53d10a3-1d16-498d-b4d8-abe7205410ec] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1185.061769] env[61995]: DEBUG oslo_concurrency.lockutils [None req-75eb1095-a0db-4c38-9379-488904743cb2 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.866s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.065430] env[61995]: DEBUG nova.network.neutron [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Updating instance_info_cache with network_info: [{"id": "4f406016-d02c-46a1-915f-44e02ed7ad90", "address": "fa:16:3e:f2:91:25", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f406016-d0", "ovs_interfaceid": "4f406016-d02c-46a1-915f-44e02ed7ad90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.309277] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.309463] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Cleaning up deleted instances with incomplete migration {{(pid=61995) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1185.334864] env[61995]: DEBUG nova.compute.manager [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Received event network-vif-plugged-4f406016-d02c-46a1-915f-44e02ed7ad90 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1185.334864] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] Acquiring lock "a3e4c652-49dc-48a1-affc-090e2b92736a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.334864] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.334864] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.334864] env[61995]: DEBUG nova.compute.manager [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] No waiting events found dispatching network-vif-plugged-4f406016-d02c-46a1-915f-44e02ed7ad90 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1185.334864] env[61995]: WARNING nova.compute.manager [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Received unexpected event network-vif-plugged-4f406016-d02c-46a1-915f-44e02ed7ad90 for instance with vm_state building and task_state spawning. [ 1185.334864] env[61995]: DEBUG nova.compute.manager [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Received event network-changed-4f406016-d02c-46a1-915f-44e02ed7ad90 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1185.334864] env[61995]: DEBUG nova.compute.manager [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Refreshing instance network info cache due to event network-changed-4f406016-d02c-46a1-915f-44e02ed7ad90. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1185.334864] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] Acquiring lock "refresh_cache-a3e4c652-49dc-48a1-affc-090e2b92736a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.567802] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-a3e4c652-49dc-48a1-affc-090e2b92736a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.568095] env[61995]: DEBUG nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Instance network_info: |[{"id": "4f406016-d02c-46a1-915f-44e02ed7ad90", "address": "fa:16:3e:f2:91:25", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f406016-d0", "ovs_interfaceid": "4f406016-d02c-46a1-915f-44e02ed7ad90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1185.568482] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] Acquired lock "refresh_cache-a3e4c652-49dc-48a1-affc-090e2b92736a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.568758] env[61995]: DEBUG nova.network.neutron [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Refreshing network info cache for port 4f406016-d02c-46a1-915f-44e02ed7ad90 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1185.570024] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:91:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aef08290-001a-4ae8-aff0-1889e2211389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f406016-d02c-46a1-915f-44e02ed7ad90', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1185.577380] env[61995]: DEBUG oslo.service.loopingcall [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1185.578329] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1185.578563] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9a7c910-3acd-424e-86f6-ca02f5f20d69 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.600182] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1185.600182] env[61995]: value = "task-795463" [ 1185.600182] env[61995]: _type = "Task" [ 1185.600182] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.607576] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795463, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.812483] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.109920] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795463, 'name': CreateVM_Task, 'duration_secs': 0.362531} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.112143] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1186.112806] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.112965] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.113404] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1186.113877] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d58630a-2b1b-4435-8744-8d335732dde1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.118376] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1186.118376] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520db05c-a51d-1d46-5dde-df1a6e2c1d62" [ 1186.118376] env[61995]: _type = "Task" [ 1186.118376] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.126342] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520db05c-a51d-1d46-5dde-df1a6e2c1d62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.273357] env[61995]: DEBUG nova.network.neutron [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Updated VIF entry in instance network info cache for port 4f406016-d02c-46a1-915f-44e02ed7ad90. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1186.273745] env[61995]: DEBUG nova.network.neutron [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Updating instance_info_cache with network_info: [{"id": "4f406016-d02c-46a1-915f-44e02ed7ad90", "address": "fa:16:3e:f2:91:25", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f406016-d0", "ovs_interfaceid": "4f406016-d02c-46a1-915f-44e02ed7ad90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.629936] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520db05c-a51d-1d46-5dde-df1a6e2c1d62, 'name': SearchDatastore_Task, 'duration_secs': 0.009941} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.630337] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.630737] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1186.630938] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.631119] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.631339] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1186.631644] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea447da9-f879-4cf7-b2a2-8a523dca8f08 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.640712] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1186.640935] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1186.641751] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3431a1b4-ad83-40d3-b89a-98b7cfe8657a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.647384] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1186.647384] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]524e2a8e-e9be-d2db-81ee-7efcaf0a4dcc" [ 1186.647384] env[61995]: _type = "Task" [ 1186.647384] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.655495] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524e2a8e-e9be-d2db-81ee-7efcaf0a4dcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.779524] env[61995]: DEBUG oslo_concurrency.lockutils [req-7b024763-2fbe-4dee-aca0-714cf08ef7d2 req-5adca2c7-2fc2-4678-bed8-00aa9c508e0d service nova] Releasing lock "refresh_cache-a3e4c652-49dc-48a1-affc-090e2b92736a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.157580] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]524e2a8e-e9be-d2db-81ee-7efcaf0a4dcc, 'name': SearchDatastore_Task, 'duration_secs': 0.00889} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.158383] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd7fa541-243c-470f-90b8-8aaff44c4a40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.163197] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1187.163197] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]529251ef-625e-ed5b-af81-2f86b34e6e2a" [ 1187.163197] env[61995]: _type = "Task" [ 1187.163197] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.170760] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529251ef-625e-ed5b-af81-2f86b34e6e2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.360967] env[61995]: DEBUG nova.compute.manager [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Received event network-changed-2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1187.361197] env[61995]: DEBUG nova.compute.manager [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Refreshing instance network info cache due to event network-changed-2a2dabdb-39a2-4193-accd-c3fb7fb129b9. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1187.361416] env[61995]: DEBUG oslo_concurrency.lockutils [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] Acquiring lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.361564] env[61995]: DEBUG oslo_concurrency.lockutils [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] Acquired lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.361731] env[61995]: DEBUG nova.network.neutron [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Refreshing network info cache for port 2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1187.672843] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]529251ef-625e-ed5b-af81-2f86b34e6e2a, 'name': SearchDatastore_Task, 'duration_secs': 0.012344} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.673062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.673325] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] a3e4c652-49dc-48a1-affc-090e2b92736a/a3e4c652-49dc-48a1-affc-090e2b92736a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1187.673575] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b97bd05-ee15-430c-af54-2604be62e720 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.680429] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1187.680429] env[61995]: value = "task-795464" [ 1187.680429] env[61995]: _type = "Task" [ 1187.680429] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.687124] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.091385] env[61995]: DEBUG nova.network.neutron [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updated VIF entry in instance network info cache for port 2a2dabdb-39a2-4193-accd-c3fb7fb129b9. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1188.091758] env[61995]: DEBUG nova.network.neutron [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updating instance_info_cache with network_info: [{"id": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "address": "fa:16:3e:3c:5e:4b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a2dabdb-39", "ovs_interfaceid": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.191643] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795464, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.594732] env[61995]: DEBUG oslo_concurrency.lockutils [req-6a3cbbdf-443d-4205-8010-ff68be982f68 req-0d7d09c4-94c3-431f-9f7e-7be5e111981b service nova] Releasing lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.690825] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538647} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.691118] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] a3e4c652-49dc-48a1-affc-090e2b92736a/a3e4c652-49dc-48a1-affc-090e2b92736a.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1188.691335] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1188.691585] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8de04db8-4252-4cbf-be76-a59e7db11fc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.700621] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1188.700621] env[61995]: value = "task-795465" [ 1188.700621] env[61995]: _type = "Task" [ 1188.700621] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.708543] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.210480] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.246273} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.210822] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1189.211475] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d488cfa-e7f5-48c4-afe3-ae6b39352e74 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.232929] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] a3e4c652-49dc-48a1-affc-090e2b92736a/a3e4c652-49dc-48a1-affc-090e2b92736a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1189.233195] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e32e1484-51e9-4842-855d-0934351c1b17 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.253317] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1189.253317] env[61995]: value = "task-795466" [ 1189.253317] env[61995]: _type = "Task" [ 1189.253317] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.260894] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795466, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.763804] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795466, 'name': ReconfigVM_Task, 'duration_secs': 0.310718} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.764116] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Reconfigured VM instance instance-00000070 to attach disk [datastore1] a3e4c652-49dc-48a1-affc-090e2b92736a/a3e4c652-49dc-48a1-affc-090e2b92736a.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1189.764747] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e96a42f-168f-4ea0-b6ba-42a6f899453e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.770881] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1189.770881] env[61995]: value = "task-795467" [ 1189.770881] env[61995]: _type = "Task" [ 1189.770881] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.778230] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795467, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.280513] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795467, 'name': Rename_Task, 'duration_secs': 0.151495} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.280806] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1190.281058] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3973ce53-9778-4977-84a0-ba37584fa64f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.287181] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1190.287181] env[61995]: value = "task-795468" [ 1190.287181] env[61995]: _type = "Task" [ 1190.287181] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.294222] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795468, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.797996] env[61995]: DEBUG oslo_vmware.api [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795468, 'name': PowerOnVM_Task, 'duration_secs': 0.440472} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.797996] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1190.798365] env[61995]: INFO nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Took 7.41 seconds to spawn the instance on the hypervisor. [ 1190.798365] env[61995]: DEBUG nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1190.799071] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e97507e-d380-46d1-8e29-18a0d6c985c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.316921] env[61995]: INFO nova.compute.manager [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Took 12.14 seconds to build instance. [ 1191.818912] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3cacb793-f071-42ff-9006-7395ee447d34 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.654s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.304287] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "a3e4c652-49dc-48a1-affc-090e2b92736a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.304586] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.808044] env[61995]: DEBUG nova.compute.utils [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1194.311187] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.178979] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.179352] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.179596] env[61995]: INFO nova.compute.manager [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Shelving [ 1195.374077] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "a3e4c652-49dc-48a1-affc-090e2b92736a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.374504] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.374627] env[61995]: INFO nova.compute.manager [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Attaching volume 7804474f-69e5-40ab-a2ff-4900fe451286 to /dev/sdb [ 1195.405401] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac936f93-964c-47b9-ba9a-606d29e01831 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.412951] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532f12ff-6e1f-420c-b54f-c657d2c650de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.426435] env[61995]: DEBUG nova.virt.block_device [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Updating existing volume attachment record: 8a3ca958-f39d-4610-a598-9c9dc1ec8023 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1195.687373] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1195.687623] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88752308-3ca7-4daa-9bd6-0dfe0068f914 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.694074] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1195.694074] env[61995]: value = "task-795472" [ 1195.694074] env[61995]: _type = "Task" [ 1195.694074] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.702455] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.203895] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795472, 'name': PowerOffVM_Task, 'duration_secs': 0.202086} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.204125] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1196.204866] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ded435-96ca-4cad-ba9e-0fca630d8c79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.222072] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38a58e1-6106-4445-88de-19db77fec9c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.732335] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1196.732812] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6141ecfc-fbb8-4cf5-8275-8a9d65238ee2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.740930] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1196.740930] env[61995]: value = "task-795473" [ 1196.740930] env[61995]: _type = "Task" [ 1196.740930] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.750629] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795473, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.250890] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795473, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.751702] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795473, 'name': CreateSnapshot_Task, 'duration_secs': 0.565614} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.752157] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1197.752769] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f12f1a-f623-419b-9c27-b20302a977f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.270398] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1198.270755] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-61d6d085-dc9a-4db4-a96a-c71c6fa6aa47 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.278903] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1198.278903] env[61995]: value = "task-795475" [ 1198.278903] env[61995]: _type = "Task" [ 1198.278903] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.286610] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795475, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.788943] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795475, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.291010] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795475, 'name': CloneVM_Task, 'duration_secs': 0.981054} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.291312] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Created linked-clone VM from snapshot [ 1199.292073] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe1bb52-9e72-4037-8807-8f07eabfa93a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.302380] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Uploading image 78580b91-0d0e-4f41-9c2b-3197fb8bb892 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1199.326643] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1199.326643] env[61995]: value = "vm-185448" [ 1199.326643] env[61995]: _type = "VirtualMachine" [ 1199.326643] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1199.326911] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-96ad9632-d2b3-4583-a37b-0b2d7adc1ee0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.335014] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease: (returnval){ [ 1199.335014] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ea7525-4943-1905-2802-3172ef4d1096" [ 1199.335014] env[61995]: _type = "HttpNfcLease" [ 1199.335014] env[61995]: } obtained for exporting VM: (result){ [ 1199.335014] env[61995]: value = "vm-185448" [ 1199.335014] env[61995]: _type = "VirtualMachine" [ 1199.335014] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1199.335306] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the lease: (returnval){ [ 1199.335306] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ea7525-4943-1905-2802-3172ef4d1096" [ 1199.335306] env[61995]: _type = "HttpNfcLease" [ 1199.335306] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1199.342627] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1199.342627] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ea7525-4943-1905-2802-3172ef4d1096" [ 1199.342627] env[61995]: _type = "HttpNfcLease" [ 1199.342627] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1199.844212] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1199.844212] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ea7525-4943-1905-2802-3172ef4d1096" [ 1199.844212] env[61995]: _type = "HttpNfcLease" [ 1199.844212] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1199.844696] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1199.844696] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ea7525-4943-1905-2802-3172ef4d1096" [ 1199.844696] env[61995]: _type = "HttpNfcLease" [ 1199.844696] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1199.845260] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3d79e6-78e9-4026-a207-9b2ffee17cfa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.852535] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5264000c-5305-82e2-cf2b-898f24367556/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1199.852715] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5264000c-5305-82e2-cf2b-898f24367556/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1199.968013] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1199.968309] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185446', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'name': 'volume-7804474f-69e5-40ab-a2ff-4900fe451286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3e4c652-49dc-48a1-affc-090e2b92736a', 'attached_at': '', 'detached_at': '', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'serial': '7804474f-69e5-40ab-a2ff-4900fe451286'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1199.969192] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829f424a-f169-4502-ac2b-c27daa944df8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.984787] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323230ff-a464-4fd2-9501-4aa103b0a4ae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.009360] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-7804474f-69e5-40ab-a2ff-4900fe451286/volume-7804474f-69e5-40ab-a2ff-4900fe451286.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1200.011661] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ee4e6e0-3831-4738-99e7-a761923e1823 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.024324] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-14da3619-9cc4-4744-b9b0-6962242c7829 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.031869] env[61995]: DEBUG oslo_vmware.api [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1200.031869] env[61995]: value = "task-795477" [ 1200.031869] env[61995]: _type = "Task" [ 1200.031869] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.040409] env[61995]: DEBUG oslo_vmware.api [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795477, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.541840] env[61995]: DEBUG oslo_vmware.api [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795477, 'name': ReconfigVM_Task, 'duration_secs': 0.34325} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.542106] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-7804474f-69e5-40ab-a2ff-4900fe451286/volume-7804474f-69e5-40ab-a2ff-4900fe451286.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1200.546765] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17a9c287-6678-42bb-b521-7c76a407811b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.562516] env[61995]: DEBUG oslo_vmware.api [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1200.562516] env[61995]: value = "task-795478" [ 1200.562516] env[61995]: _type = "Task" [ 1200.562516] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.570321] env[61995]: DEBUG oslo_vmware.api [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795478, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.073594] env[61995]: DEBUG oslo_vmware.api [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795478, 'name': ReconfigVM_Task, 'duration_secs': 0.140109} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.074039] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185446', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'name': 'volume-7804474f-69e5-40ab-a2ff-4900fe451286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3e4c652-49dc-48a1-affc-090e2b92736a', 'attached_at': '', 'detached_at': '', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'serial': '7804474f-69e5-40ab-a2ff-4900fe451286'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1202.115645] env[61995]: DEBUG nova.objects.instance [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'flavor' on Instance uuid a3e4c652-49dc-48a1-affc-090e2b92736a {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1202.620269] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c99b7332-7cc2-4b7f-90bf-9967d7f207e1 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.246s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.713435] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "a3e4c652-49dc-48a1-affc-090e2b92736a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.714355] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.714660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "a3e4c652-49dc-48a1-affc-090e2b92736a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.716050] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.716050] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.717401] env[61995]: INFO nova.compute.manager [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Terminating instance [ 1202.719536] env[61995]: DEBUG nova.compute.manager [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1202.719760] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1202.720013] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12b50105-031b-42fa-9833-4ec97d5b9469 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.727408] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1202.727408] env[61995]: value = "task-795479" [ 1202.727408] env[61995]: _type = "Task" [ 1202.727408] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.735956] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.237261] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795479, 'name': PowerOffVM_Task, 'duration_secs': 0.229939} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.237589] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1203.237747] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1203.237938] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185446', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'name': 'volume-7804474f-69e5-40ab-a2ff-4900fe451286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3e4c652-49dc-48a1-affc-090e2b92736a', 'attached_at': '', 'detached_at': '', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'serial': '7804474f-69e5-40ab-a2ff-4900fe451286'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1203.238735] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75457ed9-5ac3-436d-b593-c9a57df28fb2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.262140] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b88c0eda-a45a-402e-845d-f3476690d393 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.268911] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7476a1-12f0-4a26-ae20-48d9fa4d6527 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.289058] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c931a1d8-7886-4538-9b79-4984f31e3934 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.303404] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] The volume has not been displaced from its original location: [datastore1] volume-7804474f-69e5-40ab-a2ff-4900fe451286/volume-7804474f-69e5-40ab-a2ff-4900fe451286.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1203.311911] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1203.312337] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd65a1e8-576a-436b-a4ff-2826000a2d35 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.340737] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1203.340737] env[61995]: value = "task-795480" [ 1203.340737] env[61995]: _type = "Task" [ 1203.340737] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.348607] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795480, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.851095] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795480, 'name': ReconfigVM_Task, 'duration_secs': 0.226952} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.851429] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1203.856290] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61a865d2-f0fa-465b-a01a-4301718d5969 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.872078] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1203.872078] env[61995]: value = "task-795481" [ 1203.872078] env[61995]: _type = "Task" [ 1203.872078] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.880318] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795481, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.381479] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795481, 'name': ReconfigVM_Task, 'duration_secs': 0.395986} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.381843] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185446', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'name': 'volume-7804474f-69e5-40ab-a2ff-4900fe451286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3e4c652-49dc-48a1-affc-090e2b92736a', 'attached_at': '', 'detached_at': '', 'volume_id': '7804474f-69e5-40ab-a2ff-4900fe451286', 'serial': '7804474f-69e5-40ab-a2ff-4900fe451286'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1204.382017] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1204.382814] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a524d309-f4a0-4fda-ad95-d4a2e9e44e5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.389402] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1204.389637] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4c01bdc4-81e3-4651-833b-4a43b44fd03a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.458881] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1204.459121] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1204.459309] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleting the datastore file [datastore1] a3e4c652-49dc-48a1-affc-090e2b92736a {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1204.459578] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6bf6777-ce73-4bd6-9567-e01f8381dac7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.465757] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1204.465757] env[61995]: value = "task-795483" [ 1204.465757] env[61995]: _type = "Task" [ 1204.465757] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.473501] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795483, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.975648] env[61995]: DEBUG oslo_vmware.api [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795483, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133064} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.975920] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1204.976167] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1204.976363] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1204.976549] env[61995]: INFO nova.compute.manager [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Took 2.26 seconds to destroy the instance on the hypervisor. [ 1204.976796] env[61995]: DEBUG oslo.service.loopingcall [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1204.976987] env[61995]: DEBUG nova.compute.manager [-] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1204.977094] env[61995]: DEBUG nova.network.neutron [-] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1205.242211] env[61995]: DEBUG nova.compute.manager [req-beb40a28-eac9-408f-b972-b00f522a02c0 req-e0a45daf-eb5d-4f23-8ade-cea2a4116f23 service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Received event network-vif-deleted-4f406016-d02c-46a1-915f-44e02ed7ad90 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1205.242447] env[61995]: INFO nova.compute.manager [req-beb40a28-eac9-408f-b972-b00f522a02c0 req-e0a45daf-eb5d-4f23-8ade-cea2a4116f23 service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Neutron deleted interface 4f406016-d02c-46a1-915f-44e02ed7ad90; detaching it from the instance and deleting it from the info cache [ 1205.243051] env[61995]: DEBUG nova.network.neutron [req-beb40a28-eac9-408f-b972-b00f522a02c0 req-e0a45daf-eb5d-4f23-8ade-cea2a4116f23 service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.707852] env[61995]: DEBUG nova.network.neutron [-] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.745941] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a41df4f5-3f2d-4a8c-8c16-9b2cbd12064d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.755816] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa43773-e85c-4850-9354-3836171c6fe4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.781699] env[61995]: DEBUG nova.compute.manager [req-beb40a28-eac9-408f-b972-b00f522a02c0 req-e0a45daf-eb5d-4f23-8ade-cea2a4116f23 service nova] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Detach interface failed, port_id=4f406016-d02c-46a1-915f-44e02ed7ad90, reason: Instance a3e4c652-49dc-48a1-affc-090e2b92736a could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1206.210398] env[61995]: INFO nova.compute.manager [-] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Took 1.23 seconds to deallocate network for instance. [ 1206.762985] env[61995]: INFO nova.compute.manager [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Took 0.55 seconds to detach 1 volumes for instance. [ 1207.269692] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.269962] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.270218] env[61995]: DEBUG nova.objects.instance [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'resources' on Instance uuid a3e4c652-49dc-48a1-affc-090e2b92736a {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.838641] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b8aaf8-dfc4-4d42-9299-265bf1eb3861 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.846729] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c785c597-6236-4985-8987-40ade4b5b78d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.876413] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668ab5ae-e326-44a6-978a-c80160b9a45b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.883847] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a250226b-a09c-4639-9c68-12cda25424ee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.898222] env[61995]: DEBUG nova.compute.provider_tree [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.401915] env[61995]: DEBUG nova.scheduler.client.report [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.907054] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.928981] env[61995]: INFO nova.scheduler.client.report [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted allocations for instance a3e4c652-49dc-48a1-affc-090e2b92736a [ 1209.084094] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5264000c-5305-82e2-cf2b-898f24367556/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1209.085074] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa13343-372c-48f5-9c2e-b79c7f1edcf8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.091476] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5264000c-5305-82e2-cf2b-898f24367556/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1209.091651] env[61995]: ERROR oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5264000c-5305-82e2-cf2b-898f24367556/disk-0.vmdk due to incomplete transfer. [ 1209.091867] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0fed95c6-8be5-4a06-ad83-d172545e21c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.099404] env[61995]: DEBUG oslo_vmware.rw_handles [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5264000c-5305-82e2-cf2b-898f24367556/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1209.099644] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Uploaded image 78580b91-0d0e-4f41-9c2b-3197fb8bb892 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1209.101875] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1209.102117] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2d399398-5d18-4871-9d14-143ea8e6a590 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.108588] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1209.108588] env[61995]: value = "task-795484" [ 1209.108588] env[61995]: _type = "Task" [ 1209.108588] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.115893] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795484, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.438864] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9e49ed22-d99f-4492-a11f-c8050d2eb959 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "a3e4c652-49dc-48a1-affc-090e2b92736a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.725s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.617889] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795484, 'name': Destroy_Task, 'duration_secs': 0.492365} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.618238] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Destroyed the VM [ 1209.618564] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1209.618822] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f4f180f3-bf10-41f8-888d-c671c3ae65ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.624870] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1209.624870] env[61995]: value = "task-795485" [ 1209.624870] env[61995]: _type = "Task" [ 1209.624870] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.632008] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795485, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.134665] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795485, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.636055] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795485, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.137219] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795485, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.327488] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.327755] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.637982] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795485, 'name': RemoveSnapshot_Task, 'duration_secs': 1.624043} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.638313] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1211.638605] env[61995]: DEBUG nova.compute.manager [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1211.639403] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d32a18-81da-41a8-b58a-24a581931481 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.829754] env[61995]: DEBUG nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1212.150550] env[61995]: INFO nova.compute.manager [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Shelve offloading [ 1212.153359] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1212.153606] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35229ef3-1138-4456-93a6-5d42bb480931 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.161101] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1212.161101] env[61995]: value = "task-795486" [ 1212.161101] env[61995]: _type = "Task" [ 1212.161101] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.168631] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795486, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.352093] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.352374] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.353838] env[61995]: INFO nova.compute.claims [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1212.671880] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1212.672158] env[61995]: DEBUG nova.compute.manager [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1212.672958] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5e313f-e34f-40d4-b3ad-1a18284f7785 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.678511] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.678680] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.678864] env[61995]: DEBUG nova.network.neutron [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1213.384998] env[61995]: DEBUG nova.network.neutron [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.418365] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be020e7-9547-4c1f-bb9d-37a1d95e37e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.430508] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd6d819-331c-40b9-be99-28dfad5fb617 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.460707] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6a7fbb-e666-4956-bb80-b9e6d65f9e05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.467671] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c915bc-c5e4-469e-a989-6499facfffec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.480195] env[61995]: DEBUG nova.compute.provider_tree [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.887972] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.982555] env[61995]: DEBUG nova.scheduler.client.report [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1214.123694] env[61995]: DEBUG nova.compute.manager [req-b3af882d-4670-4677-8318-8508480fa22a req-eb98298b-0405-4f60-bfce-c170325c5580 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-vif-unplugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1214.123802] env[61995]: DEBUG oslo_concurrency.lockutils [req-b3af882d-4670-4677-8318-8508480fa22a req-eb98298b-0405-4f60-bfce-c170325c5580 service nova] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.124017] env[61995]: DEBUG oslo_concurrency.lockutils [req-b3af882d-4670-4677-8318-8508480fa22a req-eb98298b-0405-4f60-bfce-c170325c5580 service nova] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.124588] env[61995]: DEBUG oslo_concurrency.lockutils [req-b3af882d-4670-4677-8318-8508480fa22a req-eb98298b-0405-4f60-bfce-c170325c5580 service nova] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.124793] env[61995]: DEBUG nova.compute.manager [req-b3af882d-4670-4677-8318-8508480fa22a req-eb98298b-0405-4f60-bfce-c170325c5580 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] No waiting events found dispatching network-vif-unplugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1214.124980] env[61995]: WARNING nova.compute.manager [req-b3af882d-4670-4677-8318-8508480fa22a req-eb98298b-0405-4f60-bfce-c170325c5580 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received unexpected event network-vif-unplugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 for instance with vm_state shelved and task_state shelving_offloading. [ 1214.205632] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1214.206625] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b6207d-0e38-4cf1-b37d-110a79e9c957 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.214606] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1214.214835] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90b8cd30-c6f4-45f7-8159-d9b4f7e7d074 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.302552] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1214.302760] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1214.302985] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleting the datastore file [datastore2] da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1214.303269] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20c87b6d-d91d-461c-a519-2e616601ad7c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.309045] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1214.309045] env[61995]: value = "task-795488" [ 1214.309045] env[61995]: _type = "Task" [ 1214.309045] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.316482] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795488, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.487538] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.135s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.488190] env[61995]: DEBUG nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1214.818786] env[61995]: DEBUG oslo_vmware.api [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795488, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140755} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.819051] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1214.819272] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1214.819467] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1214.841279] env[61995]: INFO nova.scheduler.client.report [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted allocations for instance da5c39c9-f733-4452-9c7c-c92830682428 [ 1214.992875] env[61995]: DEBUG nova.compute.utils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1214.995059] env[61995]: DEBUG nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1214.995165] env[61995]: DEBUG nova.network.neutron [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1215.034725] env[61995]: DEBUG nova.policy [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e2f49fe94af402abe465e8415b16f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f3d88cb0eb54bc1a852c5f43f3806d9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1215.339008] env[61995]: DEBUG nova.network.neutron [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Successfully created port: 4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1215.345007] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.345298] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.345517] env[61995]: DEBUG nova.objects.instance [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'resources' on Instance uuid da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.498724] env[61995]: DEBUG nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1215.850411] env[61995]: DEBUG nova.objects.instance [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'numa_topology' on Instance uuid da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.159751] env[61995]: DEBUG nova.compute.manager [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1216.159878] env[61995]: DEBUG nova.compute.manager [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing instance network info cache due to event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1216.160116] env[61995]: DEBUG oslo_concurrency.lockutils [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.160270] env[61995]: DEBUG oslo_concurrency.lockutils [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.160440] env[61995]: DEBUG nova.network.neutron [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1216.321803] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.352559] env[61995]: DEBUG nova.objects.base [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1216.409880] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65999cdd-f108-4c61-b895-5a2b66b13b93 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.417321] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7511711b-2329-4a52-bd2e-a6598b1a8c6e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.447258] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5786b393-65a3-484a-98b2-e6d2eb57095d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.454363] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67356a2-7c04-47e6-85e4-60cb983b6416 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.468342] env[61995]: DEBUG nova.compute.provider_tree [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.507133] env[61995]: DEBUG nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1216.532461] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1216.532770] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1216.532968] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1216.533218] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1216.533396] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1216.533581] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1216.533830] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1216.534049] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1216.534273] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1216.534485] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1216.534768] env[61995]: DEBUG nova.virt.hardware [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1216.535720] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6b6496-7cbc-478e-a37b-97873863dd58 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.544206] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2bc9200-ea43-4c55-a08b-366e716380fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.802247] env[61995]: DEBUG nova.network.neutron [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Successfully updated port: 4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1216.874106] env[61995]: DEBUG nova.network.neutron [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updated VIF entry in instance network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1216.874495] env[61995]: DEBUG nova.network.neutron [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": null, "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.971538] env[61995]: DEBUG nova.scheduler.client.report [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1217.307284] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.307444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.307592] env[61995]: DEBUG nova.network.neutron [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1217.377363] env[61995]: DEBUG oslo_concurrency.lockutils [req-a72a2071-d915-498a-9313-a8f90cf5122a req-0c6bf424-2bb7-4515-8fa4-6aae7e76affa service nova] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1217.476448] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.131s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.838701] env[61995]: DEBUG nova.network.neutron [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1217.955649] env[61995]: DEBUG nova.network.neutron [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Updating instance_info_cache with network_info: [{"id": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "address": "fa:16:3e:76:37:a4", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aea4a4c-32", "ovs_interfaceid": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.983506] env[61995]: DEBUG oslo_concurrency.lockutils [None req-52d77b07-7b11-4ab4-846f-df6eeb2ddf9a tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.804s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.984227] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.663s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.984414] env[61995]: INFO nova.compute.manager [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Unshelving [ 1218.180608] env[61995]: DEBUG nova.compute.manager [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Received event network-vif-plugged-4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1218.180857] env[61995]: DEBUG oslo_concurrency.lockutils [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] Acquiring lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.181041] env[61995]: DEBUG oslo_concurrency.lockutils [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.181226] env[61995]: DEBUG oslo_concurrency.lockutils [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.181403] env[61995]: DEBUG nova.compute.manager [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] No waiting events found dispatching network-vif-plugged-4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1218.181575] env[61995]: WARNING nova.compute.manager [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Received unexpected event network-vif-plugged-4aea4a4c-32e7-4bf9-90dd-7397133429e1 for instance with vm_state building and task_state spawning. [ 1218.181736] env[61995]: DEBUG nova.compute.manager [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Received event network-changed-4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1218.181911] env[61995]: DEBUG nova.compute.manager [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Refreshing instance network info cache due to event network-changed-4aea4a4c-32e7-4bf9-90dd-7397133429e1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1218.182094] env[61995]: DEBUG oslo_concurrency.lockutils [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] Acquiring lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.458539] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.458860] env[61995]: DEBUG nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Instance network_info: |[{"id": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "address": "fa:16:3e:76:37:a4", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aea4a4c-32", "ovs_interfaceid": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1218.459231] env[61995]: DEBUG oslo_concurrency.lockutils [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] Acquired lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.459428] env[61995]: DEBUG nova.network.neutron [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Refreshing network info cache for port 4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1218.460682] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:37:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aef08290-001a-4ae8-aff0-1889e2211389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4aea4a4c-32e7-4bf9-90dd-7397133429e1', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1218.468037] env[61995]: DEBUG oslo.service.loopingcall [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1218.470737] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1218.471200] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e042f69-9c46-4ece-b485-9f7b77416974 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.492019] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1218.492019] env[61995]: value = "task-795489" [ 1218.492019] env[61995]: _type = "Task" [ 1218.492019] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.502327] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795489, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.746181] env[61995]: DEBUG nova.network.neutron [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Updated VIF entry in instance network info cache for port 4aea4a4c-32e7-4bf9-90dd-7397133429e1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1218.746600] env[61995]: DEBUG nova.network.neutron [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Updating instance_info_cache with network_info: [{"id": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "address": "fa:16:3e:76:37:a4", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aea4a4c-32", "ovs_interfaceid": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.004311] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795489, 'name': CreateVM_Task, 'duration_secs': 0.30488} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.004560] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1219.005134] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.005307] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.005638] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1219.005883] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-251058b0-a923-4ebb-bbc7-f9bf1b89a8be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.010478] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1219.010478] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]527f843e-280d-0078-abf7-f902ebac3fe6" [ 1219.010478] env[61995]: _type = "Task" [ 1219.010478] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.014496] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.014741] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.014959] env[61995]: DEBUG nova.objects.instance [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'pci_requests' on Instance uuid da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.020723] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527f843e-280d-0078-abf7-f902ebac3fe6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.249805] env[61995]: DEBUG oslo_concurrency.lockutils [req-37f97cf8-f1c9-47a7-a06f-df2b954ced12 req-358adbd6-022c-4011-b3e7-1146cb6dbdc5 service nova] Releasing lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.518804] env[61995]: DEBUG nova.objects.instance [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'numa_topology' on Instance uuid da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.523122] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]527f843e-280d-0078-abf7-f902ebac3fe6, 'name': SearchDatastore_Task, 'duration_secs': 0.011266} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.523599] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.523829] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1219.524069] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.524224] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.524409] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1219.524713] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3ab1aec-3dee-4e4d-9c5e-480a467ecffb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.532877] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1219.533052] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1219.533716] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7c98ca0-51d0-4e10-8598-6519003d3bda {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.538633] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1219.538633] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52e8ef14-70cb-054c-bcb8-4c5184a156e6" [ 1219.538633] env[61995]: _type = "Task" [ 1219.538633] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.545994] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e8ef14-70cb-054c-bcb8-4c5184a156e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.024486] env[61995]: INFO nova.compute.claims [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1220.051834] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52e8ef14-70cb-054c-bcb8-4c5184a156e6, 'name': SearchDatastore_Task, 'duration_secs': 0.008566} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.052630] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9faa13b9-afee-45a2-ab76-ecac7d1df02e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.058151] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1220.058151] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522d9357-b042-f09f-8f74-f7c4eb8066ea" [ 1220.058151] env[61995]: _type = "Task" [ 1220.058151] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.067066] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522d9357-b042-f09f-8f74-f7c4eb8066ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.567726] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522d9357-b042-f09f-8f74-f7c4eb8066ea, 'name': SearchDatastore_Task, 'duration_secs': 0.010177} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.568008] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.568320] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0/31bd74c4-6a7d-4760-8334-7c9ec0fa11d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1220.568585] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1864d6ab-01b1-4d67-bde8-1bad4160718e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.575405] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1220.575405] env[61995]: value = "task-795490" [ 1220.575405] env[61995]: _type = "Task" [ 1220.575405] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.583072] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.093956] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795490, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.126768] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb80349-851d-4199-bdba-a5ff12bcdf37 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.135736] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65187846-7de0-426c-8549-8b5eab509c56 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.170471] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f41cce-bedc-483c-9f02-be5ae2a7abd9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.179833] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090130cf-de8e-4500-bc69-7f65a4c855f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.193460] env[61995]: DEBUG nova.compute.provider_tree [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.586540] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634515} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.586810] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0/31bd74c4-6a7d-4760-8334-7c9ec0fa11d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1221.587041] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1221.587300] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e32f5cdf-b67a-4cc6-abd7-385249f5e071 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.593651] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1221.593651] env[61995]: value = "task-795491" [ 1221.593651] env[61995]: _type = "Task" [ 1221.593651] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.601371] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.697178] env[61995]: DEBUG nova.scheduler.client.report [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1222.104418] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093699} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.104748] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1222.105476] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d3fc6d-a4ae-43da-8f40-a10e54e85e63 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.127680] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0/31bd74c4-6a7d-4760-8334-7c9ec0fa11d0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1222.127825] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea24fee2-7ef4-4882-ba3c-5a6d4e86d667 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.146507] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1222.146507] env[61995]: value = "task-795492" [ 1222.146507] env[61995]: _type = "Task" [ 1222.146507] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.154290] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795492, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.202669] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.187s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.231268] env[61995]: INFO nova.network.neutron [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating port fef53ce7-5db4-47e6-a55a-7da9bff3cda2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1222.386467] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.386884] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.387177] env[61995]: INFO nova.compute.manager [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Shelving [ 1222.589959] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.590270] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.590464] env[61995]: DEBUG nova.compute.manager [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1222.591398] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa4b23e-b49d-459e-9a3a-1ab8a7ff5150 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.598207] env[61995]: DEBUG nova.compute.manager [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1222.598803] env[61995]: DEBUG nova.objects.instance [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'flavor' on Instance uuid c99b0295-bf1f-4a25-af31-cd0e19b6b089 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.656248] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795492, 'name': ReconfigVM_Task, 'duration_secs': 0.283819} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.656534] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0/31bd74c4-6a7d-4760-8334-7c9ec0fa11d0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1222.657156] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b86507f4-084d-43c7-a85f-69605a27ed3b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.663384] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1222.663384] env[61995]: value = "task-795493" [ 1222.663384] env[61995]: _type = "Task" [ 1222.663384] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.670888] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.895090] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1222.895090] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9802be14-da1f-4c32-9eec-336048d5a573 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.901697] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1222.901697] env[61995]: value = "task-795494" [ 1222.901697] env[61995]: _type = "Task" [ 1222.901697] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.909207] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795494, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.103901] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1223.104234] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f487061-fff6-480f-a487-b3493947da52 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.111842] env[61995]: DEBUG oslo_vmware.api [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1223.111842] env[61995]: value = "task-795495" [ 1223.111842] env[61995]: _type = "Task" [ 1223.111842] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.120242] env[61995]: DEBUG oslo_vmware.api [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.173224] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.411012] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795494, 'name': PowerOffVM_Task, 'duration_secs': 0.209215} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.411334] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1223.412098] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6007c0-d846-4f22-9d6b-abed946e7ece {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.430758] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38eb60c6-5b94-4935-b92d-984d8d8e8488 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.625244] env[61995]: DEBUG oslo_vmware.api [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795495, 'name': PowerOffVM_Task, 'duration_secs': 0.196943} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.625244] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1223.625244] env[61995]: DEBUG nova.compute.manager [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1223.625244] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b735863-d995-4520-aea5-7f716a320c6b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.657166] env[61995]: DEBUG nova.compute.manager [req-d2516873-56d7-48d3-b2f4-861f8c664368 req-95ed3336-2e62-42dd-8001-5204a1a811b8 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-vif-plugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1223.657166] env[61995]: DEBUG oslo_concurrency.lockutils [req-d2516873-56d7-48d3-b2f4-861f8c664368 req-95ed3336-2e62-42dd-8001-5204a1a811b8 service nova] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.657166] env[61995]: DEBUG oslo_concurrency.lockutils [req-d2516873-56d7-48d3-b2f4-861f8c664368 req-95ed3336-2e62-42dd-8001-5204a1a811b8 service nova] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.657166] env[61995]: DEBUG oslo_concurrency.lockutils [req-d2516873-56d7-48d3-b2f4-861f8c664368 req-95ed3336-2e62-42dd-8001-5204a1a811b8 service nova] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.657166] env[61995]: DEBUG nova.compute.manager [req-d2516873-56d7-48d3-b2f4-861f8c664368 req-95ed3336-2e62-42dd-8001-5204a1a811b8 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] No waiting events found dispatching network-vif-plugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1223.657166] env[61995]: WARNING nova.compute.manager [req-d2516873-56d7-48d3-b2f4-861f8c664368 req-95ed3336-2e62-42dd-8001-5204a1a811b8 service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received unexpected event network-vif-plugged-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 for instance with vm_state shelved_offloaded and task_state spawning. [ 1223.674371] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.716371] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1223.716578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1223.716716] env[61995]: DEBUG nova.network.neutron [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1223.940354] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1223.940767] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-80d74018-f1ab-4333-bbc5-6fb30cbb3732 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.948789] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1223.948789] env[61995]: value = "task-795496" [ 1223.948789] env[61995]: _type = "Task" [ 1223.948789] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.957533] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795496, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.135918] env[61995]: DEBUG oslo_concurrency.lockutils [None req-ed6e6829-6cd3-4b76-8ab8-8bd8397e07d5 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.174509] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.459396] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795496, 'name': CreateSnapshot_Task, 'duration_secs': 0.437183} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.459753] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1224.460641] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44333eb8-9132-42ff-b04e-f1414319b1b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.675521] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.729745] env[61995]: DEBUG nova.network.neutron [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.978970] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1224.979320] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e86f5d62-e67c-4c31-9b24-f7ffcb6efd01 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.987770] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1224.987770] env[61995]: value = "task-795497" [ 1224.987770] env[61995]: _type = "Task" [ 1224.987770] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.997102] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795497, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.006298] env[61995]: DEBUG nova.objects.instance [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'flavor' on Instance uuid c99b0295-bf1f-4a25-af31-cd0e19b6b089 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.175216] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.232202] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1225.258842] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5bb8ca7ed2e89e523324d90de787f45e',container_format='bare',created_at=2024-09-10T09:04:06Z,direct_url=,disk_format='vmdk',id=78580b91-0d0e-4f41-9c2b-3197fb8bb892,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-139306723-shelved',owner='cd23f8abd8f14ec392fbfb7fd5bc64f9',properties=ImageMetaProps,protected=,size=31661056,status='active',tags=,updated_at=2024-09-10T09:04:20Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1225.259165] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1225.259337] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1225.259528] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1225.259677] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1225.259836] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1225.260064] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1225.260254] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1225.260456] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1225.260616] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1225.260794] env[61995]: DEBUG nova.virt.hardware [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1225.261654] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e209de-2794-40c7-bc25-c06d2dde2a94 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.269943] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe14efd-e2c0-4b69-b29b-2db2084627d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.284643] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:d3:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7894814c-6be3-4b80-a08e-4a771bc05dd1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fef53ce7-5db4-47e6-a55a-7da9bff3cda2', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1225.292157] env[61995]: DEBUG oslo.service.loopingcall [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1225.292431] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1225.292649] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d2f8016-4c9c-49d1-ac4f-d9de8fa42abd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.312350] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1225.312350] env[61995]: value = "task-795498" [ 1225.312350] env[61995]: _type = "Task" [ 1225.312350] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.321481] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795498, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.497820] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795497, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.512845] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.513057] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.513247] env[61995]: DEBUG nova.network.neutron [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1225.513433] env[61995]: DEBUG nova.objects.instance [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'info_cache' on Instance uuid c99b0295-bf1f-4a25-af31-cd0e19b6b089 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.676010] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.681801] env[61995]: DEBUG nova.compute.manager [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1225.682029] env[61995]: DEBUG nova.compute.manager [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing instance network info cache due to event network-changed-fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1225.682227] env[61995]: DEBUG oslo_concurrency.lockutils [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.682378] env[61995]: DEBUG oslo_concurrency.lockutils [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.682542] env[61995]: DEBUG nova.network.neutron [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Refreshing network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1225.822143] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795498, 'name': CreateVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.999924] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795497, 'name': CloneVM_Task, 'duration_secs': 0.988329} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.000217] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Created linked-clone VM from snapshot [ 1226.000915] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdfe3a5-9bd8-4d88-82ca-8cc0b371b495 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.007720] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Uploading image 6a74f8d6-3b45-497c-92f2-1c2a3019f865 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1226.016632] env[61995]: DEBUG nova.objects.base [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1226.028300] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1226.028300] env[61995]: value = "vm-185451" [ 1226.028300] env[61995]: _type = "VirtualMachine" [ 1226.028300] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1226.028746] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3fc31380-0629-4d68-b859-2630836e964a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.034838] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease: (returnval){ [ 1226.034838] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5245dba0-511c-cd70-0e08-be38d11c8a0c" [ 1226.034838] env[61995]: _type = "HttpNfcLease" [ 1226.034838] env[61995]: } obtained for exporting VM: (result){ [ 1226.034838] env[61995]: value = "vm-185451" [ 1226.034838] env[61995]: _type = "VirtualMachine" [ 1226.034838] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1226.035076] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the lease: (returnval){ [ 1226.035076] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5245dba0-511c-cd70-0e08-be38d11c8a0c" [ 1226.035076] env[61995]: _type = "HttpNfcLease" [ 1226.035076] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1226.040996] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1226.040996] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5245dba0-511c-cd70-0e08-be38d11c8a0c" [ 1226.040996] env[61995]: _type = "HttpNfcLease" [ 1226.040996] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1226.176965] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795493, 'name': Rename_Task, 'duration_secs': 3.026377} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.177244] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1226.177352] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aea4975e-ebb1-4f15-aaee-b754d278bf60 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.183396] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1226.183396] env[61995]: value = "task-795500" [ 1226.183396] env[61995]: _type = "Task" [ 1226.183396] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.192300] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795500, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.324719] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795498, 'name': CreateVM_Task, 'duration_secs': 0.546898} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.324883] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1226.325496] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1226.325668] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.326084] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1226.326337] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2050e53-cfd0-4499-b86b-5efbd5d0a8c5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.330738] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1226.330738] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d0b197-289c-8293-da68-05968bbc3a37" [ 1226.330738] env[61995]: _type = "Task" [ 1226.330738] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.338335] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d0b197-289c-8293-da68-05968bbc3a37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.389534] env[61995]: DEBUG nova.network.neutron [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updated VIF entry in instance network info cache for port fef53ce7-5db4-47e6-a55a-7da9bff3cda2. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1226.389904] env[61995]: DEBUG nova.network.neutron [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [{"id": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "address": "fa:16:3e:cf:d3:cf", "network": {"id": "24b0a46c-b126-4f67-9636-71cc9e503ff0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-959378744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd23f8abd8f14ec392fbfb7fd5bc64f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfef53ce7-5d", "ovs_interfaceid": "fef53ce7-5db4-47e6-a55a-7da9bff3cda2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.542906] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1226.542906] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5245dba0-511c-cd70-0e08-be38d11c8a0c" [ 1226.542906] env[61995]: _type = "HttpNfcLease" [ 1226.542906] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1226.543220] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1226.543220] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5245dba0-511c-cd70-0e08-be38d11c8a0c" [ 1226.543220] env[61995]: _type = "HttpNfcLease" [ 1226.543220] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1226.543914] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9de3f3-65bb-462d-afa0-90f37d5412b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.552516] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d80212-12f8-77e1-d7ca-c78c3994de6d/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1226.552696] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d80212-12f8-77e1-d7ca-c78c3994de6d/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1226.643553] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b8b3b25d-820d-4532-afa3-53b9327a5d80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.695249] env[61995]: DEBUG oslo_vmware.api [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795500, 'name': PowerOnVM_Task, 'duration_secs': 0.485218} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.695522] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1226.695726] env[61995]: INFO nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Took 10.19 seconds to spawn the instance on the hypervisor. [ 1226.695903] env[61995]: DEBUG nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1226.696745] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc63378-2f14-4c20-bf9a-1c8635f5cc62 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.786529] env[61995]: DEBUG nova.network.neutron [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updating instance_info_cache with network_info: [{"id": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "address": "fa:16:3e:3c:5e:4b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a2dabdb-39", "ovs_interfaceid": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.841623] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1226.841904] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Processing image 78580b91-0d0e-4f41-9c2b-3197fb8bb892 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1226.842159] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892/78580b91-0d0e-4f41-9c2b-3197fb8bb892.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1226.842312] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892/78580b91-0d0e-4f41-9c2b-3197fb8bb892.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.842494] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1226.842734] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c24cb53d-ac22-4934-b00f-c1a7a109ea61 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.856218] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1226.856433] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1226.857144] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4690f54-74ba-4413-88f8-aede21102205 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.862344] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1226.862344] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522684e1-1dff-08da-6963-81877d60c0aa" [ 1226.862344] env[61995]: _type = "Task" [ 1226.862344] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.869709] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522684e1-1dff-08da-6963-81877d60c0aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.892500] env[61995]: DEBUG oslo_concurrency.lockutils [req-f70d99c4-b87c-4068-9c88-69bd9eea0387 req-85669e6e-bb34-4840-9de9-0877d8a803dd service nova] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.214495] env[61995]: INFO nova.compute.manager [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Took 14.88 seconds to build instance. [ 1227.289014] env[61995]: DEBUG oslo_concurrency.lockutils [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.373869] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Preparing fetch location {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1227.374409] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Fetch image to [datastore1] OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163/OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163.vmdk {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1227.375117] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Downloading stream optimized image 78580b91-0d0e-4f41-9c2b-3197fb8bb892 to [datastore1] OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163/OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163.vmdk on the data store datastore1 as vApp {{(pid=61995) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1227.375117] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Downloading image file data 78580b91-0d0e-4f41-9c2b-3197fb8bb892 to the ESX as VM named 'OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163' {{(pid=61995) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1227.676273] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1227.676273] env[61995]: value = "resgroup-9" [ 1227.676273] env[61995]: _type = "ResourcePool" [ 1227.676273] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1227.676660] env[61995]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-12045d9c-127c-491b-898d-b4ab0b50ad73 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.698221] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease: (returnval){ [ 1227.698221] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b69a65-00c7-dc15-f0b5-8a25469cb695" [ 1227.698221] env[61995]: _type = "HttpNfcLease" [ 1227.698221] env[61995]: } obtained for vApp import into resource pool (val){ [ 1227.698221] env[61995]: value = "resgroup-9" [ 1227.698221] env[61995]: _type = "ResourcePool" [ 1227.698221] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1227.699381] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the lease: (returnval){ [ 1227.699381] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b69a65-00c7-dc15-f0b5-8a25469cb695" [ 1227.699381] env[61995]: _type = "HttpNfcLease" [ 1227.699381] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1227.705702] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1227.705702] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b69a65-00c7-dc15-f0b5-8a25469cb695" [ 1227.705702] env[61995]: _type = "HttpNfcLease" [ 1227.705702] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1227.716963] env[61995]: DEBUG oslo_concurrency.lockutils [None req-80ee353d-aac9-4b4a-8631-a68d7c6fbc02 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.389s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.792222] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1227.792616] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb7ee3f8-f859-4db5-ba27-dfb75924de9f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.799936] env[61995]: DEBUG oslo_vmware.api [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1227.799936] env[61995]: value = "task-795502" [ 1227.799936] env[61995]: _type = "Task" [ 1227.799936] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.808477] env[61995]: DEBUG oslo_vmware.api [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795502, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.153958] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.154294] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.154596] env[61995]: INFO nova.compute.manager [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Shelving [ 1228.207618] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1228.207618] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b69a65-00c7-dc15-f0b5-8a25469cb695" [ 1228.207618] env[61995]: _type = "HttpNfcLease" [ 1228.207618] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1228.310277] env[61995]: DEBUG oslo_vmware.api [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795502, 'name': PowerOnVM_Task, 'duration_secs': 0.50785} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.310277] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1228.312024] env[61995]: DEBUG nova.compute.manager [None req-91c17a66-c5e0-42b2-be7c-3093d3b27991 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1228.312348] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f8dd76-11f0-4169-a119-c4486487be7a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.663566] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1228.663749] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca462d21-0d7e-4ae8-bf4d-346549e37cc9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.672439] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1228.672439] env[61995]: value = "task-795503" [ 1228.672439] env[61995]: _type = "Task" [ 1228.672439] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.681713] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795503, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.709881] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1228.709881] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b69a65-00c7-dc15-f0b5-8a25469cb695" [ 1228.709881] env[61995]: _type = "HttpNfcLease" [ 1228.709881] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1228.710293] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1228.710293] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b69a65-00c7-dc15-f0b5-8a25469cb695" [ 1228.710293] env[61995]: _type = "HttpNfcLease" [ 1228.710293] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1228.711026] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75a6ba5-61b2-4402-8b65-f23af1ee5688 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.720549] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ccfa88-085b-35d6-3101-561a0c24d555/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1228.720777] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating HTTP connection to write to file with size = 31661056 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ccfa88-085b-35d6-3101-561a0c24d555/disk-0.vmdk. {{(pid=61995) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1228.784438] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f7013164-313b-4687-baed-266e297ac16f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.185780] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795503, 'name': PowerOffVM_Task, 'duration_secs': 0.304036} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.188693] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1229.189780] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76761d6f-7d5e-44c7-b516-3e65aaeec5eb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.216885] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42834a1c-1911-4f1a-8ac4-7ab02610e0e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.731205] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1229.732061] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3035113d-6737-4978-a553-eec47c0a4ad3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.741253] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1229.741253] env[61995]: value = "task-795504" [ 1229.741253] env[61995]: _type = "Task" [ 1229.741253] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.751555] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795504, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.986255] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Completed reading data from the image iterator. {{(pid=61995) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1229.986587] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ccfa88-085b-35d6-3101-561a0c24d555/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1229.987814] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d55059-7f1c-4bac-96a7-f83387500540 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.995370] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ccfa88-085b-35d6-3101-561a0c24d555/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1229.995605] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ccfa88-085b-35d6-3101-561a0c24d555/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1229.995885] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b5430ad3-7ef6-4588-8a49-17663c783ee0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.133239] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2cbd54a-f060-40d0-8eb6-840a18b0de05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.140760] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1d273fc4-438f-4fba-b064-50fccdaeed9f tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Suspending the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1230.141059] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4c2d08f6-02c3-488d-8474-427631ab6833 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.147430] env[61995]: DEBUG oslo_vmware.api [None req-1d273fc4-438f-4fba-b064-50fccdaeed9f tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1230.147430] env[61995]: value = "task-795505" [ 1230.147430] env[61995]: _type = "Task" [ 1230.147430] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.155480] env[61995]: DEBUG oslo_vmware.api [None req-1d273fc4-438f-4fba-b064-50fccdaeed9f tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795505, 'name': SuspendVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.212050] env[61995]: DEBUG oslo_vmware.rw_handles [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ccfa88-085b-35d6-3101-561a0c24d555/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1230.212333] env[61995]: INFO nova.virt.vmwareapi.images [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Downloaded image file data 78580b91-0d0e-4f41-9c2b-3197fb8bb892 [ 1230.213178] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95cd1ec-10b4-48f8-aeba-7ca39191a17b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.228708] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a62fc78f-161d-498e-ad38-5bc0082784fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.252778] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795504, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.254445] env[61995]: INFO nova.virt.vmwareapi.images [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] The imported VM was unregistered [ 1230.256368] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Caching image {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1230.256617] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Creating directory with path [datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1230.256899] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f7a173e1-7cb1-4066-8aa2-456be02dc9c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.268212] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Created directory with path [datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1230.268413] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163/OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163.vmdk to [datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892/78580b91-0d0e-4f41-9c2b-3197fb8bb892.vmdk. {{(pid=61995) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1230.268731] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-deddb54b-9318-4a78-a58d-aec238e426ce {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.276425] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1230.276425] env[61995]: value = "task-795507" [ 1230.276425] env[61995]: _type = "Task" [ 1230.276425] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.284589] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795507, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.659811] env[61995]: DEBUG oslo_vmware.api [None req-1d273fc4-438f-4fba-b064-50fccdaeed9f tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795505, 'name': SuspendVM_Task} progress is 62%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.754878] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795504, 'name': CreateSnapshot_Task, 'duration_secs': 0.824794} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.755209] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1230.756031] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d435ce8a-bed2-40a6-894c-84c888756de8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.787942] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795507, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.161798] env[61995]: DEBUG oslo_vmware.api [None req-1d273fc4-438f-4fba-b064-50fccdaeed9f tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795505, 'name': SuspendVM_Task, 'duration_secs': 0.661198} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.162101] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1d273fc4-438f-4fba-b064-50fccdaeed9f tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Suspended the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1231.162301] env[61995]: DEBUG nova.compute.manager [None req-1d273fc4-438f-4fba-b064-50fccdaeed9f tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1231.163132] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ffdfa31-9497-4e3e-96dd-6722e850e7ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.277538] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1231.278463] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a673bcf2-0b46-4a98-aa19-eda69dbbdb16 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.290536] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795507, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.292709] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1231.292709] env[61995]: value = "task-795508" [ 1231.292709] env[61995]: _type = "Task" [ 1231.292709] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.302886] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795508, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.791627] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795507, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.805725] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795508, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.293382] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795507, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.306891] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795508, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.661742] env[61995]: INFO nova.compute.manager [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Resuming [ 1232.662400] env[61995]: DEBUG nova.objects.instance [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'flavor' on Instance uuid c99b0295-bf1f-4a25-af31-cd0e19b6b089 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.789415] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795507, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.311792} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.789675] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163/OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163.vmdk to [datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892/78580b91-0d0e-4f41-9c2b-3197fb8bb892.vmdk. [ 1232.789868] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Cleaning up location [datastore1] OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1232.790048] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_09990129-8d50-451e-b586-3cb61ca42163 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1232.790302] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e053c01-c299-43f2-a5f2-8eb23c44b542 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.799439] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1232.799439] env[61995]: value = "task-795509" [ 1232.799439] env[61995]: _type = "Task" [ 1232.799439] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.805976] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795508, 'name': CloneVM_Task, 'duration_secs': 1.461006} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.806599] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Created linked-clone VM from snapshot [ 1232.807353] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b1ab33-ac8e-48a6-bb63-825af1c507cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.812297] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795509, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.816910] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Uploading image 2de0c526-f0b7-412e-a18f-3184cbe8a130 {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1232.842200] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1232.842200] env[61995]: value = "vm-185455" [ 1232.842200] env[61995]: _type = "VirtualMachine" [ 1232.842200] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1232.842514] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8e9c3bf3-0d73-4278-b97e-364437622e1e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.849710] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lease: (returnval){ [ 1232.849710] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292fa75-235c-379d-39ca-aea565998df5" [ 1232.849710] env[61995]: _type = "HttpNfcLease" [ 1232.849710] env[61995]: } obtained for exporting VM: (result){ [ 1232.849710] env[61995]: value = "vm-185455" [ 1232.849710] env[61995]: _type = "VirtualMachine" [ 1232.849710] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1232.850025] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the lease: (returnval){ [ 1232.850025] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292fa75-235c-379d-39ca-aea565998df5" [ 1232.850025] env[61995]: _type = "HttpNfcLease" [ 1232.850025] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1232.856580] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1232.856580] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292fa75-235c-379d-39ca-aea565998df5" [ 1232.856580] env[61995]: _type = "HttpNfcLease" [ 1232.856580] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1233.309945] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795509, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.04766} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.310250] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1233.310525] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892/78580b91-0d0e-4f41-9c2b-3197fb8bb892.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1233.310793] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892/78580b91-0d0e-4f41-9c2b-3197fb8bb892.vmdk to [datastore1] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1233.311063] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b19e485a-abac-47f3-89d7-58d785f938c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.317789] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1233.317789] env[61995]: value = "task-795511" [ 1233.317789] env[61995]: _type = "Task" [ 1233.317789] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.325718] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795511, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.357868] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1233.357868] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292fa75-235c-379d-39ca-aea565998df5" [ 1233.357868] env[61995]: _type = "HttpNfcLease" [ 1233.357868] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1233.358209] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1233.358209] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5292fa75-235c-379d-39ca-aea565998df5" [ 1233.358209] env[61995]: _type = "HttpNfcLease" [ 1233.358209] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1233.358954] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747209d1-b86a-4cd3-9be8-57df9ad10af2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.366233] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6d44a-3981-20e1-fa8e-55ed41157a0e/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1233.366424] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6d44a-3981-20e1-fa8e-55ed41157a0e/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1233.457566] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7432b7a0-c0a2-44ae-9a48-05451ca16bc9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.670903] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1233.671157] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquired lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.671256] env[61995]: DEBUG nova.network.neutron [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1233.828931] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795511, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.333265] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795511, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.502129] env[61995]: DEBUG nova.network.neutron [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updating instance_info_cache with network_info: [{"id": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "address": "fa:16:3e:3c:5e:4b", "network": {"id": "31703010-5e86-4661-b23f-4a11092c3e0d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1952919046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c3e03b3825c43ca987d60b339fc0a52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ada35c98-01a9-4352-98e4-1d20ba31f928", "external-id": "nsx-vlan-transportzone-242", "segmentation_id": 242, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a2dabdb-39", "ovs_interfaceid": "2a2dabdb-39a2-4193-accd-c3fb7fb129b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1234.831129] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795511, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.006385] env[61995]: DEBUG oslo_concurrency.lockutils [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Releasing lock "refresh_cache-c99b0295-bf1f-4a25-af31-cd0e19b6b089" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.007631] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4129002-fe66-4ad3-ae48-d06c2acd6b5f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.015100] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Resuming the VM {{(pid=61995) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1235.015406] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cfbb655-7c96-4393-b583-fd2d1eb73b0e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.023212] env[61995]: DEBUG oslo_vmware.api [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1235.023212] env[61995]: value = "task-795512" [ 1235.023212] env[61995]: _type = "Task" [ 1235.023212] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.033395] env[61995]: DEBUG oslo_vmware.api [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795512, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.332865] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795511, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.537584] env[61995]: DEBUG oslo_vmware.api [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795512, 'name': PowerOnVM_Task} progress is 93%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.834594] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795511, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.309989} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.835138] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/78580b91-0d0e-4f41-9c2b-3197fb8bb892/78580b91-0d0e-4f41-9c2b-3197fb8bb892.vmdk to [datastore1] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1235.835787] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d473ec-9414-4e32-9c21-4091a94e60b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.861980] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1235.862360] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d306dd6c-7691-4ad2-b793-59cfc10eabe4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.885286] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1235.885286] env[61995]: value = "task-795513" [ 1235.885286] env[61995]: _type = "Task" [ 1235.885286] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.895805] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795513, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.036191] env[61995]: DEBUG oslo_vmware.api [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795512, 'name': PowerOnVM_Task, 'duration_secs': 0.631392} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.036555] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Resumed the VM {{(pid=61995) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1236.036803] env[61995]: DEBUG nova.compute.manager [None req-4692a975-c8d5-4613-8549-2d5f4ff16047 tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1236.037725] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33b527f-17cf-47d7-b451-4478f37b616c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.255476] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d80212-12f8-77e1-d7ca-c78c3994de6d/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1236.256442] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29f5b3e-6756-43d2-a0e4-ecd91cb3dd65 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.263202] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d80212-12f8-77e1-d7ca-c78c3994de6d/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1236.263400] env[61995]: ERROR oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d80212-12f8-77e1-d7ca-c78c3994de6d/disk-0.vmdk due to incomplete transfer. [ 1236.263625] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7315fffd-ee4f-493d-92e4-6b42fbbf275f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.271742] env[61995]: DEBUG oslo_vmware.rw_handles [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d80212-12f8-77e1-d7ca-c78c3994de6d/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1236.271944] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Uploaded image 6a74f8d6-3b45-497c-92f2-1c2a3019f865 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1236.274338] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1236.274618] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-405e2d8f-c80e-4ee9-be90-5655e0144675 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.280395] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1236.280395] env[61995]: value = "task-795514" [ 1236.280395] env[61995]: _type = "Task" [ 1236.280395] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.288580] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795514, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.394853] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795513, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.790378] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795514, 'name': Destroy_Task, 'duration_secs': 0.339304} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.790592] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Destroyed the VM [ 1236.790833] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1236.791097] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0cf89b59-894a-4cfc-8028-5397446ce8f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.797701] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1236.797701] env[61995]: value = "task-795515" [ 1236.797701] env[61995]: _type = "Task" [ 1236.797701] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.807333] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795515, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.895294] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795513, 'name': ReconfigVM_Task, 'duration_secs': 0.527528} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.895756] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Reconfigured VM instance instance-0000006c to attach disk [datastore1] da5c39c9-f733-4452-9c7c-c92830682428/da5c39c9-f733-4452-9c7c-c92830682428.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1236.896307] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd20cf58-1d8d-4d0d-81c5-0b4fb5367833 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.903139] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1236.903139] env[61995]: value = "task-795516" [ 1236.903139] env[61995]: _type = "Task" [ 1236.903139] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.911266] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795516, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.308112] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795515, 'name': RemoveSnapshot_Task, 'duration_secs': 0.366603} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.308429] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1237.308678] env[61995]: DEBUG nova.compute.manager [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1237.309444] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8d3510-2042-46e7-8994-a55e78f2a305 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.412479] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795516, 'name': Rename_Task, 'duration_secs': 0.14852} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.412759] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1237.412976] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4bd5ce0a-89df-4361-8dac-ea87f483cc0a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.419362] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1237.419362] env[61995]: value = "task-795517" [ 1237.419362] env[61995]: _type = "Task" [ 1237.419362] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.426824] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795517, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.481718] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.482018] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.482266] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.482461] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.482639] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.484921] env[61995]: INFO nova.compute.manager [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Terminating instance [ 1237.486808] env[61995]: DEBUG nova.compute.manager [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1237.487015] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1237.487868] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ce15be-d12b-4144-9e87-01800c0f2336 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.495240] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1237.495474] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8cbecb7-bc1a-4c6d-b3b6-ed2bb0335fb1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.501266] env[61995]: DEBUG oslo_vmware.api [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1237.501266] env[61995]: value = "task-795518" [ 1237.501266] env[61995]: _type = "Task" [ 1237.501266] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.508929] env[61995]: DEBUG oslo_vmware.api [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795518, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.822504] env[61995]: INFO nova.compute.manager [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Shelve offloading [ 1237.824442] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1237.824709] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f1322b3-9be1-4d15-9c05-61a68219f46c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.833253] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1237.833253] env[61995]: value = "task-795519" [ 1237.833253] env[61995]: _type = "Task" [ 1237.833253] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.842811] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1237.843060] env[61995]: DEBUG nova.compute.manager [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1237.843888] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af17bdc7-bb54-4c05-877a-b454913e19fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.850149] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1237.850363] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.850580] env[61995]: DEBUG nova.network.neutron [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1237.929774] env[61995]: DEBUG oslo_vmware.api [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795517, 'name': PowerOnVM_Task, 'duration_secs': 0.476914} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.930287] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1238.011302] env[61995]: DEBUG oslo_vmware.api [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795518, 'name': PowerOffVM_Task, 'duration_secs': 0.21655} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.011614] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1238.011757] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1238.012015] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-318f5429-80f3-4c0f-bfc9-fa83c13494e6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.031637] env[61995]: DEBUG nova.compute.manager [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1238.032695] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eca6118-f352-4da0-a454-66d16a506a51 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.508024] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1238.508024] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1238.508024] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleting the datastore file [datastore2] c99b0295-bf1f-4a25-af31-cd0e19b6b089 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1238.508388] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-190768d7-3151-4365-937b-ba33b0df5b16 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.514475] env[61995]: DEBUG oslo_vmware.api [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for the task: (returnval){ [ 1238.514475] env[61995]: value = "task-795521" [ 1238.514475] env[61995]: _type = "Task" [ 1238.514475] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.522940] env[61995]: DEBUG oslo_vmware.api [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795521, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.549837] env[61995]: DEBUG oslo_concurrency.lockutils [None req-470eaaea-235f-480a-b743-26a17c4d2de3 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.566s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.624365] env[61995]: DEBUG nova.network.neutron [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47ddce51-cf", "ovs_interfaceid": "47ddce51-cfec-49c4-ba3b-b34c86140652", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.024719] env[61995]: DEBUG oslo_vmware.api [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Task: {'id': task-795521, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316766} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.024992] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1239.025202] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1239.025384] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1239.025564] env[61995]: INFO nova.compute.manager [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Took 1.54 seconds to destroy the instance on the hypervisor. [ 1239.025818] env[61995]: DEBUG oslo.service.loopingcall [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1239.026019] env[61995]: DEBUG nova.compute.manager [-] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1239.026843] env[61995]: DEBUG nova.network.neutron [-] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1239.127618] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.435979] env[61995]: DEBUG nova.compute.manager [req-68f59820-4f40-48f4-a255-771186af0bc3 req-dc08ec02-573e-451d-bf05-a961d1c9f013 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-vif-unplugged-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1239.435979] env[61995]: DEBUG oslo_concurrency.lockutils [req-68f59820-4f40-48f4-a255-771186af0bc3 req-dc08ec02-573e-451d-bf05-a961d1c9f013 service nova] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.435979] env[61995]: DEBUG oslo_concurrency.lockutils [req-68f59820-4f40-48f4-a255-771186af0bc3 req-dc08ec02-573e-451d-bf05-a961d1c9f013 service nova] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.435979] env[61995]: DEBUG oslo_concurrency.lockutils [req-68f59820-4f40-48f4-a255-771186af0bc3 req-dc08ec02-573e-451d-bf05-a961d1c9f013 service nova] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.435979] env[61995]: DEBUG nova.compute.manager [req-68f59820-4f40-48f4-a255-771186af0bc3 req-dc08ec02-573e-451d-bf05-a961d1c9f013 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] No waiting events found dispatching network-vif-unplugged-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1239.435979] env[61995]: WARNING nova.compute.manager [req-68f59820-4f40-48f4-a255-771186af0bc3 req-dc08ec02-573e-451d-bf05-a961d1c9f013 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received unexpected event network-vif-unplugged-47ddce51-cfec-49c4-ba3b-b34c86140652 for instance with vm_state shelved and task_state shelving_offloading. [ 1239.519696] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1239.521483] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab75cbce-986d-494c-99ae-81a885b6677b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.530715] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1239.530976] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51d3aadd-d893-4e82-8771-e1ec780ddc6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.601979] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1239.601979] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1239.601979] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleting the datastore file [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1239.601979] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8eb9145-a54a-4e5d-aa92-aa9c4654e53d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.607099] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1239.607099] env[61995]: value = "task-795523" [ 1239.607099] env[61995]: _type = "Task" [ 1239.607099] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.617756] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795523, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.639578] env[61995]: DEBUG nova.compute.manager [req-3c9d24d1-782e-47cd-8081-24ac66b28711 req-da6dd619-a37c-4896-91ad-a585d2248ead service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Received event network-vif-deleted-2a2dabdb-39a2-4193-accd-c3fb7fb129b9 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1239.639965] env[61995]: INFO nova.compute.manager [req-3c9d24d1-782e-47cd-8081-24ac66b28711 req-da6dd619-a37c-4896-91ad-a585d2248ead service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Neutron deleted interface 2a2dabdb-39a2-4193-accd-c3fb7fb129b9; detaching it from the instance and deleting it from the info cache [ 1239.640038] env[61995]: DEBUG nova.network.neutron [req-3c9d24d1-782e-47cd-8081-24ac66b28711 req-da6dd619-a37c-4896-91ad-a585d2248ead service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.901580] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.901975] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.902231] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "da5c39c9-f733-4452-9c7c-c92830682428-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.902435] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.902731] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.905227] env[61995]: INFO nova.compute.manager [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Terminating instance [ 1239.907210] env[61995]: DEBUG nova.compute.manager [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1239.907417] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1239.908382] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bfa34e-3e66-4e0f-8bae-44b918d8b28e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.916893] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1239.917170] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53e69279-ca85-4f3c-a8c7-fa4ada4bd719 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.923275] env[61995]: DEBUG oslo_vmware.api [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1239.923275] env[61995]: value = "task-795524" [ 1239.923275] env[61995]: _type = "Task" [ 1239.923275] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.932200] env[61995]: DEBUG oslo_vmware.api [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.116031] env[61995]: DEBUG oslo_vmware.api [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795523, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174186} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.116397] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1240.116508] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1240.116689] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1240.118976] env[61995]: DEBUG nova.network.neutron [-] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.142963] env[61995]: INFO nova.scheduler.client.report [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted allocations for instance 24853688-b972-48e6-be7b-7f0cf634c60d [ 1240.148474] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-325a51fe-9e1f-4394-8e2d-143fb4fae3b1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.157428] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648020c5-55f9-415e-8301-ecfeb6a16cd7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.188986] env[61995]: DEBUG nova.compute.manager [req-3c9d24d1-782e-47cd-8081-24ac66b28711 req-da6dd619-a37c-4896-91ad-a585d2248ead service nova] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Detach interface failed, port_id=2a2dabdb-39a2-4193-accd-c3fb7fb129b9, reason: Instance c99b0295-bf1f-4a25-af31-cd0e19b6b089 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1240.433517] env[61995]: DEBUG oslo_vmware.api [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795524, 'name': PowerOffVM_Task, 'duration_secs': 0.200225} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.433794] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1240.433964] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1240.434232] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ce00016-0c4a-4472-bef8-31fbbe20a3ec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.623114] env[61995]: INFO nova.compute.manager [-] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Took 1.60 seconds to deallocate network for instance. [ 1240.643784] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1240.644015] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1240.644219] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleting the datastore file [datastore1] da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1240.644492] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2349fcd8-0598-4c52-a435-8649fce7490d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.650046] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.650046] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.650253] env[61995]: DEBUG nova.objects.instance [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'resources' on Instance uuid 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1240.653363] env[61995]: DEBUG oslo_vmware.api [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for the task: (returnval){ [ 1240.653363] env[61995]: value = "task-795526" [ 1240.653363] env[61995]: _type = "Task" [ 1240.653363] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.663539] env[61995]: DEBUG oslo_vmware.api [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.129408] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.155324] env[61995]: DEBUG nova.objects.instance [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'numa_topology' on Instance uuid 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1241.166353] env[61995]: DEBUG oslo_vmware.api [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Task: {'id': task-795526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152467} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.166599] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1241.166821] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1241.167045] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1241.167247] env[61995]: INFO nova.compute.manager [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Took 1.26 seconds to destroy the instance on the hypervisor. [ 1241.167520] env[61995]: DEBUG oslo.service.loopingcall [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1241.167719] env[61995]: DEBUG nova.compute.manager [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1241.167816] env[61995]: DEBUG nova.network.neutron [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1241.563477] env[61995]: DEBUG nova.compute.manager [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1241.563690] env[61995]: DEBUG nova.compute.manager [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing instance network info cache due to event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1241.563965] env[61995]: DEBUG oslo_concurrency.lockutils [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] Acquiring lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.565823] env[61995]: DEBUG oslo_concurrency.lockutils [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] Acquired lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.565823] env[61995]: DEBUG nova.network.neutron [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1241.660985] env[61995]: DEBUG nova.objects.base [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Object Instance<24853688-b972-48e6-be7b-7f0cf634c60d> lazy-loaded attributes: resources,numa_topology {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1241.724136] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6abecf7-c2c4-4571-ab4c-48b213da0f76 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.732718] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2350202-d3c2-4976-8957-d92e9f8bea5d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.765120] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f3470d-3037-4855-a8be-3b94dcd73204 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.773103] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08785fc0-5549-444b-b3f1-e7ee61f4434a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.787976] env[61995]: DEBUG nova.compute.provider_tree [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1241.805329] env[61995]: DEBUG nova.compute.manager [req-603188b0-e040-426d-8566-990bbe0bf1b2 req-e44dd131-61a8-49d2-bbf9-3c7805d7a63f service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Received event network-vif-deleted-fef53ce7-5db4-47e6-a55a-7da9bff3cda2 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1241.805656] env[61995]: INFO nova.compute.manager [req-603188b0-e040-426d-8566-990bbe0bf1b2 req-e44dd131-61a8-49d2-bbf9-3c7805d7a63f service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Neutron deleted interface fef53ce7-5db4-47e6-a55a-7da9bff3cda2; detaching it from the instance and deleting it from the info cache [ 1241.805760] env[61995]: DEBUG nova.network.neutron [req-603188b0-e040-426d-8566-990bbe0bf1b2 req-e44dd131-61a8-49d2-bbf9-3c7805d7a63f service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.288494] env[61995]: DEBUG nova.network.neutron [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updated VIF entry in instance network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1242.290252] env[61995]: DEBUG nova.network.neutron [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap47ddce51-cf", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.295298] env[61995]: DEBUG nova.network.neutron [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.309109] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e743c4ca-942a-4d0f-ac5a-5238cfbe68e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.313946] env[61995]: ERROR nova.scheduler.client.report [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [req-fdcbdb09-666b-42c0-b8df-1c4867d97983] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fdcbdb09-666b-42c0-b8df-1c4867d97983"}]} [ 1242.320058] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0f136e-2180-495a-b6e9-31e4355e237a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.332475] env[61995]: DEBUG nova.scheduler.client.report [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1242.351916] env[61995]: DEBUG nova.compute.manager [req-603188b0-e040-426d-8566-990bbe0bf1b2 req-e44dd131-61a8-49d2-bbf9-3c7805d7a63f service nova] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Detach interface failed, port_id=fef53ce7-5db4-47e6-a55a-7da9bff3cda2, reason: Instance da5c39c9-f733-4452-9c7c-c92830682428 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1242.353393] env[61995]: DEBUG nova.scheduler.client.report [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1242.353393] env[61995]: DEBUG nova.compute.provider_tree [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1242.364625] env[61995]: DEBUG nova.scheduler.client.report [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1242.385112] env[61995]: DEBUG nova.scheduler.client.report [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1242.411985] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6d44a-3981-20e1-fa8e-55ed41157a0e/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1242.413135] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e644831c-a22f-4561-875d-fde20055c8f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.420449] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6d44a-3981-20e1-fa8e-55ed41157a0e/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1242.420677] env[61995]: ERROR oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6d44a-3981-20e1-fa8e-55ed41157a0e/disk-0.vmdk due to incomplete transfer. [ 1242.420931] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-caf80c5f-5f9d-462f-b48f-313c4c5f4eed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.431260] env[61995]: DEBUG oslo_vmware.rw_handles [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6d44a-3981-20e1-fa8e-55ed41157a0e/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1242.431260] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Uploaded image 2de0c526-f0b7-412e-a18f-3184cbe8a130 to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1242.433070] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1242.433521] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-419d6c38-71ed-435a-b8dd-f69189e28131 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.442280] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1242.442280] env[61995]: value = "task-795527" [ 1242.442280] env[61995]: _type = "Task" [ 1242.442280] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.450597] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795527, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.452607] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961d6896-aad2-4fcf-a86e-0b9d6bfc287e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.459244] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89033151-7a5b-489c-b6f3-5cc6f717b6fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.489565] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d029242-929b-445e-8f54-5d9e5a083609 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.497462] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be79c179-9b25-4398-95e3-e03f1f07f88c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.511672] env[61995]: DEBUG nova.compute.provider_tree [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1242.796216] env[61995]: DEBUG oslo_concurrency.lockutils [req-3d174e15-db8e-4100-b286-ba246688d32c req-235d5a09-139a-4218-8460-9aac72c8bd74 service nova] Releasing lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.797812] env[61995]: INFO nova.compute.manager [-] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Took 1.63 seconds to deallocate network for instance. [ 1242.911016] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.952055] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795527, 'name': Destroy_Task, 'duration_secs': 0.338035} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.952350] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Destroyed the VM [ 1242.952587] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1242.952838] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d8a703c6-2469-4dd2-b51f-1df3dfcfcd05 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.959502] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1242.959502] env[61995]: value = "task-795528" [ 1242.959502] env[61995]: _type = "Task" [ 1242.959502] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.966964] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795528, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.044434] env[61995]: DEBUG nova.scheduler.client.report [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 159 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1243.044742] env[61995]: DEBUG nova.compute.provider_tree [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 159 to 160 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1243.044946] env[61995]: DEBUG nova.compute.provider_tree [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1243.304853] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.469523] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795528, 'name': RemoveSnapshot_Task, 'duration_secs': 0.377281} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.469776] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1243.470065] env[61995]: DEBUG nova.compute.manager [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1243.470856] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e42489-a4cd-4d08-830d-15c346ca63b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.550016] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.900s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.552866] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.423s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.552866] env[61995]: DEBUG nova.objects.instance [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lazy-loading 'resources' on Instance uuid c99b0295-bf1f-4a25-af31-cd0e19b6b089 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1243.982584] env[61995]: INFO nova.compute.manager [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Shelve offloading [ 1243.984289] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1243.984533] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d4acc48-9c1a-42a3-bf63-cfdca46adb79 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.992546] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1243.992546] env[61995]: value = "task-795529" [ 1243.992546] env[61995]: _type = "Task" [ 1243.992546] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.000386] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.059877] env[61995]: DEBUG oslo_concurrency.lockutils [None req-88c462b4-2371-49d5-b615-09a8379939a7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.673s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.060760] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.150s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.061101] env[61995]: INFO nova.compute.manager [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Unshelving [ 1244.113383] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d33d9c-4dc0-4286-8aca-62c45b71b3a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.120789] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3e075d-52a5-4ed9-bedc-5ceb77b4953d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.149725] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252da7f4-f03b-42a1-9c6a-5f7df5eac760 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.156459] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db2c8f4-cd24-42dd-86e6-6fb6a3d96da3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.169951] env[61995]: DEBUG nova.compute.provider_tree [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1244.503886] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1244.504269] env[61995]: DEBUG nova.compute.manager [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1244.504871] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b115d8-ded0-446a-bbd0-cbf9e7719cfe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.510839] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.511025] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.511199] env[61995]: DEBUG nova.network.neutron [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1244.673117] env[61995]: DEBUG nova.scheduler.client.report [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1245.072329] env[61995]: DEBUG nova.compute.utils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1245.178195] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.180367] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.876s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.180563] env[61995]: DEBUG nova.objects.instance [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lazy-loading 'resources' on Instance uuid da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1245.200226] env[61995]: INFO nova.scheduler.client.report [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Deleted allocations for instance c99b0295-bf1f-4a25-af31-cd0e19b6b089 [ 1245.288015] env[61995]: DEBUG nova.network.neutron [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Updating instance_info_cache with network_info: [{"id": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "address": "fa:16:3e:76:37:a4", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aea4a4c-32", "ovs_interfaceid": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.575124] env[61995]: INFO nova.virt.block_device [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Booting with volume 9834a3bf-0fe3-4814-8d27-5aadd4a878ed at /dev/sdb [ 1245.609542] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52e34898-40be-442e-877c-58dd1c6b8a18 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.618360] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ced4023-cd73-4668-a1f1-1b5cef2a23e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.640800] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dbbb4c2b-2b8f-4412-95e9-e5c0f1e33364 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.648136] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5548d4-a3ad-4662-b9a9-d759a3f09cb5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.669844] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa02f843-3e01-42c5-9dc4-f44f2d8bed2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.675489] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d9fe65-2502-4667-8e4b-dc1aaff0feb1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.690330] env[61995]: DEBUG nova.virt.block_device [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating existing volume attachment record: e696a5fa-c189-42ef-8fbb-91cc54576083 {{(pid=61995) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1245.707329] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e5cda602-25a9-46d9-b829-63b400aac23e tempest-ServerActionsTestJSON-2147235425 tempest-ServerActionsTestJSON-2147235425-project-member] Lock "c99b0295-bf1f-4a25-af31-cd0e19b6b089" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.225s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.736097] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae29ef2d-90e4-4788-ad27-44466cbff6fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.743852] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2654376-9587-41b6-8986-52c4bbb90ac9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.772220] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c28bfe-4725-4e5a-82cb-0258f54753db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.778703] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731ac2bc-eee7-41c2-ac2e-021d9e2973e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.791086] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.792926] env[61995]: DEBUG nova.compute.provider_tree [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1246.043971] env[61995]: DEBUG nova.compute.manager [req-acbe252e-416a-4436-8982-ef903dbbf566 req-4aa3dc67-290d-4579-9a58-468392443f13 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Received event network-vif-unplugged-4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1246.044207] env[61995]: DEBUG oslo_concurrency.lockutils [req-acbe252e-416a-4436-8982-ef903dbbf566 req-4aa3dc67-290d-4579-9a58-468392443f13 service nova] Acquiring lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.044793] env[61995]: DEBUG oslo_concurrency.lockutils [req-acbe252e-416a-4436-8982-ef903dbbf566 req-4aa3dc67-290d-4579-9a58-468392443f13 service nova] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.044793] env[61995]: DEBUG oslo_concurrency.lockutils [req-acbe252e-416a-4436-8982-ef903dbbf566 req-4aa3dc67-290d-4579-9a58-468392443f13 service nova] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1246.044793] env[61995]: DEBUG nova.compute.manager [req-acbe252e-416a-4436-8982-ef903dbbf566 req-4aa3dc67-290d-4579-9a58-468392443f13 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] No waiting events found dispatching network-vif-unplugged-4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1246.044976] env[61995]: WARNING nova.compute.manager [req-acbe252e-416a-4436-8982-ef903dbbf566 req-4aa3dc67-290d-4579-9a58-468392443f13 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Received unexpected event network-vif-unplugged-4aea4a4c-32e7-4bf9-90dd-7397133429e1 for instance with vm_state shelved and task_state shelving_offloading. [ 1246.069200] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1246.070464] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fece07bf-5c6b-45ad-abbb-2d5c9939dd90 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.079221] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1246.079449] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6901b97c-26e3-4c13-8be8-62fd22da43a0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.150493] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1246.150790] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1246.150997] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleting the datastore file [datastore1] 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1246.151285] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b1e2e4d-4c24-480e-b1db-11a14a5e6287 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.157622] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1246.157622] env[61995]: value = "task-795534" [ 1246.157622] env[61995]: _type = "Task" [ 1246.157622] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.166086] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.296514] env[61995]: DEBUG nova.scheduler.client.report [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1246.315795] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.315964] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.316378] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1246.316562] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1246.667459] env[61995]: DEBUG oslo_vmware.api [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152185} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.667878] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1246.667878] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1246.668074] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1246.690766] env[61995]: INFO nova.scheduler.client.report [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted allocations for instance 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0 [ 1246.801418] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1246.817380] env[61995]: INFO nova.scheduler.client.report [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Deleted allocations for instance da5c39c9-f733-4452-9c7c-c92830682428 [ 1246.837409] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1246.837553] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1246.837690] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1246.837839] env[61995]: DEBUG nova.objects.instance [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lazy-loading 'info_cache' on Instance uuid da5c39c9-f733-4452-9c7c-c92830682428 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.195760] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.196084] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.196360] env[61995]: DEBUG nova.objects.instance [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'resources' on Instance uuid 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.327186] env[61995]: DEBUG oslo_concurrency.lockutils [None req-e50604cd-4bcc-4298-b473-7941174cde13 tempest-ServerActionsTestOtherB-977073590 tempest-ServerActionsTestOtherB-977073590-project-member] Lock "da5c39c9-f733-4452-9c7c-c92830682428" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.425s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.698663] env[61995]: DEBUG nova.objects.instance [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'numa_topology' on Instance uuid 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.843022] env[61995]: DEBUG nova.compute.utils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Can not refresh info_cache because instance was not found {{(pid=61995) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 1247.858733] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1247.920136] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.071917] env[61995]: DEBUG nova.compute.manager [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Received event network-changed-4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1248.072107] env[61995]: DEBUG nova.compute.manager [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Refreshing instance network info cache due to event network-changed-4aea4a4c-32e7-4bf9-90dd-7397133429e1. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1248.072351] env[61995]: DEBUG oslo_concurrency.lockutils [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] Acquiring lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.072510] env[61995]: DEBUG oslo_concurrency.lockutils [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] Acquired lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.072670] env[61995]: DEBUG nova.network.neutron [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Refreshing network info cache for port 4aea4a4c-32e7-4bf9-90dd-7397133429e1 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1248.209958] env[61995]: DEBUG nova.objects.base [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Object Instance<31bd74c4-6a7d-4760-8334-7c9ec0fa11d0> lazy-loaded attributes: resources,numa_topology {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1248.244384] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b29fc5-f091-4246-9496-e47223950d15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.252406] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d74c62-8afa-4f1f-bb10-f7f28c53f140 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.283331] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c075af-0184-42a7-aebb-7bee3e9fe129 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.291073] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5cd431-ee24-4692-8e8c-e5898d0917c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.305054] env[61995]: DEBUG nova.compute.provider_tree [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.441652] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1248.811168] env[61995]: DEBUG nova.scheduler.client.report [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1248.835752] env[61995]: DEBUG nova.network.neutron [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Updated VIF entry in instance network info cache for port 4aea4a4c-32e7-4bf9-90dd-7397133429e1. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1248.836134] env[61995]: DEBUG nova.network.neutron [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Updating instance_info_cache with network_info: [{"id": "4aea4a4c-32e7-4bf9-90dd-7397133429e1", "address": "fa:16:3e:76:37:a4", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": null, "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4aea4a4c-32", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1248.944379] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-da5c39c9-f733-4452-9c7c-c92830682428" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1248.944547] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1248.944744] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.944901] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.945058] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.945234] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.945399] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.945546] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.945673] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1248.945813] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.314488] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.118s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.338135] env[61995]: DEBUG oslo_concurrency.lockutils [req-676515e1-18c6-46f5-b186-8369b6c446e3 req-aee174be-7034-4a73-b1a7-6673fba76045 service nova] Releasing lock "refresh_cache-31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1249.448280] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.448547] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.448719] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.448871] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1249.449857] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b342ffe-9cdc-46b4-a898-461f94a295b0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.458746] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baca4ad3-1ff2-4d30-ab5f-90ee5abbbe46 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.474618] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f1d2ee-4ea7-430b-ba77-1133e5d850cc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.482177] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ac5f2c-c7ea-4ff2-ae40-1e4c3e7a5cb6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.516411] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181261MB free_disk=67GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1249.516704] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.516958] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.825980] env[61995]: DEBUG oslo_concurrency.lockutils [None req-64d22bfe-7d43-4778-be97-9d0c9fd837ab tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.672s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.827291] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.907s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.827577] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.827825] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.828053] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.832621] env[61995]: INFO nova.compute.manager [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Terminating instance [ 1249.838706] env[61995]: DEBUG nova.compute.manager [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1249.838706] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1249.838706] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b8b29b2-0a16-4655-a610-c0b2066449a8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.847759] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9605811-6883-434a-978b-d57322deac0f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.873289] env[61995]: WARNING nova.virt.vmwareapi.vmops [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0 could not be found. [ 1249.874862] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1249.875190] env[61995]: INFO nova.compute.manager [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1249.875558] env[61995]: DEBUG oslo.service.loopingcall [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1249.875864] env[61995]: DEBUG nova.compute.manager [-] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1249.876071] env[61995]: DEBUG nova.network.neutron [-] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1250.627060] env[61995]: DEBUG nova.network.neutron [-] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.051022] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 24853688-b972-48e6-be7b-7f0cf634c60d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1251.051022] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1251.051022] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1251.090320] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c6115e-18ca-4a2c-a8ea-2462d349f43e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.098378] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d4aa3f-3254-4471-916e-f607c7a7e288 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.127813] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39cda83-0fb7-438a-b730-cb2245e7a4e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.130692] env[61995]: INFO nova.compute.manager [-] [instance: 31bd74c4-6a7d-4760-8334-7c9ec0fa11d0] Took 1.25 seconds to deallocate network for instance. [ 1251.138844] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2161e222-e391-439b-a38f-aad5508ebbca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.154464] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.304587] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.462495] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "87efc871-d14b-4b9f-98c0-87474651cbff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.462732] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "87efc871-d14b-4b9f-98c0-87474651cbff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.655476] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1251.964604] env[61995]: DEBUG nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1252.160487] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9b6318e7-603c-432e-a167-ca780bdb45f2 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "31bd74c4-6a7d-4760-8334-7c9ec0fa11d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.333s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.164784] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1252.164979] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.648s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.165271] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.862s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.165516] env[61995]: DEBUG nova.objects.instance [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'pci_requests' on Instance uuid 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1252.491673] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.670645] env[61995]: DEBUG nova.objects.instance [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'numa_topology' on Instance uuid 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.172660] env[61995]: INFO nova.compute.claims [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1253.275465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.275683] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.777765] env[61995]: DEBUG nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1254.249415] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd232ce-8b13-4369-8170-8957e0e0f88d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.258797] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c62bb6-955f-4450-87ee-ac4ac0aa3f08 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.293926] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a5a2c1-4a37-4be0-92bd-ab3aea5ba361 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.301758] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99012d2-74a2-4cd4-ae7e-6eb316123987 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.317775] env[61995]: DEBUG nova.compute.provider_tree [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.319758] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.821971] env[61995]: DEBUG nova.scheduler.client.report [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1255.328596] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.163s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.330764] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.839s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.333029] env[61995]: INFO nova.compute.claims [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1255.366720] env[61995]: INFO nova.network.neutron [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating port 47ddce51-cfec-49c4-ba3b-b34c86140652 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1255.717023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.717023] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.221067] env[61995]: DEBUG nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1256.402127] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5e5714-8ba8-4175-ae11-2fe3d3f52f0c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.411121] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b8d46f-0394-418e-8b34-b09f850bafcf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.440717] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c23e1a-1267-4618-8eb7-eef3b417639a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.448085] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7aea94-a5ea-4fbe-a705-ad9b1bbb88a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.460707] env[61995]: DEBUG nova.compute.provider_tree [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1256.737682] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.963872] env[61995]: DEBUG nova.scheduler.client.report [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1257.317611] env[61995]: DEBUG nova.compute.manager [req-c34225ee-f909-4cb0-a8f0-44f0028a242e req-f69692d8-73e4-454c-a6a0-0951accadb5f service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-vif-plugged-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1257.317845] env[61995]: DEBUG oslo_concurrency.lockutils [req-c34225ee-f909-4cb0-a8f0-44f0028a242e req-f69692d8-73e4-454c-a6a0-0951accadb5f service nova] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.318102] env[61995]: DEBUG oslo_concurrency.lockutils [req-c34225ee-f909-4cb0-a8f0-44f0028a242e req-f69692d8-73e4-454c-a6a0-0951accadb5f service nova] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.318314] env[61995]: DEBUG oslo_concurrency.lockutils [req-c34225ee-f909-4cb0-a8f0-44f0028a242e req-f69692d8-73e4-454c-a6a0-0951accadb5f service nova] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.318493] env[61995]: DEBUG nova.compute.manager [req-c34225ee-f909-4cb0-a8f0-44f0028a242e req-f69692d8-73e4-454c-a6a0-0951accadb5f service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] No waiting events found dispatching network-vif-plugged-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1257.318662] env[61995]: WARNING nova.compute.manager [req-c34225ee-f909-4cb0-a8f0-44f0028a242e req-f69692d8-73e4-454c-a6a0-0951accadb5f service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received unexpected event network-vif-plugged-47ddce51-cfec-49c4-ba3b-b34c86140652 for instance with vm_state shelved_offloaded and task_state spawning. [ 1257.399727] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.399924] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.400116] env[61995]: DEBUG nova.network.neutron [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1257.469200] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.138s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.469719] env[61995]: DEBUG nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1257.472204] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.152s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.473557] env[61995]: INFO nova.compute.claims [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1257.977999] env[61995]: DEBUG nova.compute.utils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1257.982466] env[61995]: DEBUG nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1257.986150] env[61995]: DEBUG nova.network.neutron [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1258.025476] env[61995]: DEBUG nova.policy [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97b5fdaed2b247139796e43c62940dc7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c706e3ad98f7431b891f4df1fffc7745', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1258.168448] env[61995]: DEBUG nova.network.neutron [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47ddce51-cf", "ovs_interfaceid": "47ddce51-cfec-49c4-ba3b-b34c86140652", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1258.311297] env[61995]: DEBUG nova.network.neutron [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Successfully created port: 82531ec7-7339-4d0d-9535-099ebf8ee9b7 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1258.492345] env[61995]: DEBUG nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1258.561868] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9895c3a-36a0-4446-9fe0-d66b64314108 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.569828] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393e1566-4ab3-4da2-9029-2da9cddada20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.598446] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f2a922-ba53-4f1d-b3b6-c8595ad3772b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.605576] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c17408-e057-410f-b273-fc8648c30868 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.618374] env[61995]: DEBUG nova.compute.provider_tree [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1258.671125] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.699647] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1d755902528ea3d4751d384e5efe099c',container_format='bare',created_at=2024-09-10T09:04:33Z,direct_url=,disk_format='vmdk',id=6a74f8d6-3b45-497c-92f2-1c2a3019f865,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-378797491-shelved',owner='49e78af0ad9340258211bf92e447021c',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-09-10T09:04:48Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1258.699943] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1258.700130] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1258.700388] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1258.700614] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1258.700783] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1258.701069] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1258.701336] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1258.701603] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1258.701798] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1258.701981] env[61995]: DEBUG nova.virt.hardware [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1258.702831] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8647dc-9742-49f0-85b1-ca06e33289d3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.710673] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ff56d3-2317-475e-bff5-34145b60bddd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.724996] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:74:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '28d04eee-6dbb-491a-a999-b659c799679d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47ddce51-cfec-49c4-ba3b-b34c86140652', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1258.732120] env[61995]: DEBUG oslo.service.loopingcall [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1258.732356] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1258.732915] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-72e8a349-b50c-486d-b48e-ade2afd50228 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.751816] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1258.751816] env[61995]: value = "task-795537" [ 1258.751816] env[61995]: _type = "Task" [ 1258.751816] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.759328] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795537, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.122066] env[61995]: DEBUG nova.scheduler.client.report [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1259.262059] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795537, 'name': CreateVM_Task, 'duration_secs': 0.381925} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.262270] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1259.262906] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1259.263087] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.263470] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1259.263727] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6744e0b1-a2a3-46fe-bb7c-6b60db91e728 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.268586] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1259.268586] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]526c7024-56bd-3faf-8d44-f0feebac855e" [ 1259.268586] env[61995]: _type = "Task" [ 1259.268586] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.275697] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]526c7024-56bd-3faf-8d44-f0feebac855e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.342091] env[61995]: DEBUG nova.compute.manager [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1259.342372] env[61995]: DEBUG nova.compute.manager [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing instance network info cache due to event network-changed-47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1259.342673] env[61995]: DEBUG oslo_concurrency.lockutils [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] Acquiring lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1259.342845] env[61995]: DEBUG oslo_concurrency.lockutils [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] Acquired lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.343089] env[61995]: DEBUG nova.network.neutron [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Refreshing network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1259.505059] env[61995]: DEBUG nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1259.527176] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1259.527481] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1259.527685] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1259.527914] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1259.528116] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1259.528341] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1259.528587] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1259.528788] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1259.528995] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1259.529205] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1259.529421] env[61995]: DEBUG nova.virt.hardware [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1259.530374] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94c4955-f96d-417b-a2ba-966f84ca69fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.538812] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d98ee3-08c6-45ad-8e61-8b1cfe53a677 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.625946] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.154s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.626586] env[61995]: DEBUG nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1259.629102] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.892s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1259.631102] env[61995]: INFO nova.compute.claims [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1259.730754] env[61995]: DEBUG nova.compute.manager [req-d3adebf7-500d-4561-ab97-966752c072ec req-68b66adb-b5dd-49d7-81ae-a3bbf94a767a service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Received event network-vif-plugged-82531ec7-7339-4d0d-9535-099ebf8ee9b7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1259.731070] env[61995]: DEBUG oslo_concurrency.lockutils [req-d3adebf7-500d-4561-ab97-966752c072ec req-68b66adb-b5dd-49d7-81ae-a3bbf94a767a service nova] Acquiring lock "87efc871-d14b-4b9f-98c0-87474651cbff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1259.731333] env[61995]: DEBUG oslo_concurrency.lockutils [req-d3adebf7-500d-4561-ab97-966752c072ec req-68b66adb-b5dd-49d7-81ae-a3bbf94a767a service nova] Lock "87efc871-d14b-4b9f-98c0-87474651cbff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1259.731531] env[61995]: DEBUG oslo_concurrency.lockutils [req-d3adebf7-500d-4561-ab97-966752c072ec req-68b66adb-b5dd-49d7-81ae-a3bbf94a767a service nova] Lock "87efc871-d14b-4b9f-98c0-87474651cbff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.731744] env[61995]: DEBUG nova.compute.manager [req-d3adebf7-500d-4561-ab97-966752c072ec req-68b66adb-b5dd-49d7-81ae-a3bbf94a767a service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] No waiting events found dispatching network-vif-plugged-82531ec7-7339-4d0d-9535-099ebf8ee9b7 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1259.732074] env[61995]: WARNING nova.compute.manager [req-d3adebf7-500d-4561-ab97-966752c072ec req-68b66adb-b5dd-49d7-81ae-a3bbf94a767a service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Received unexpected event network-vif-plugged-82531ec7-7339-4d0d-9535-099ebf8ee9b7 for instance with vm_state building and task_state spawning. [ 1259.778868] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.779100] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Processing image 6a74f8d6-3b45-497c-92f2-1c2a3019f865 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1259.779354] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865/6a74f8d6-3b45-497c-92f2-1c2a3019f865.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1259.779503] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865/6a74f8d6-3b45-497c-92f2-1c2a3019f865.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.779684] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1259.779922] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20a4668a-ec84-4db2-be18-862526b3ec29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.789761] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1259.789945] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1259.790667] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0709a415-bd68-47d2-a5bb-2bbb824d883b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.795268] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1259.795268] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52491934-7798-749e-2442-d99826ac4c84" [ 1259.795268] env[61995]: _type = "Task" [ 1259.795268] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.480322] env[61995]: DEBUG nova.network.neutron [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Successfully updated port: 82531ec7-7339-4d0d-9535-099ebf8ee9b7 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1260.484419] env[61995]: DEBUG nova.compute.utils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1260.488062] env[61995]: DEBUG nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1260.488260] env[61995]: DEBUG nova.network.neutron [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1260.494739] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Preparing fetch location {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1260.494985] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Fetch image to [datastore2] OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d/OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d.vmdk {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1260.495183] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Downloading stream optimized image 6a74f8d6-3b45-497c-92f2-1c2a3019f865 to [datastore2] OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d/OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d.vmdk on the data store datastore2 as vApp {{(pid=61995) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1260.495358] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Downloading image file data 6a74f8d6-3b45-497c-92f2-1c2a3019f865 to the ESX as VM named 'OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d' {{(pid=61995) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1260.541540] env[61995]: DEBUG nova.policy [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e2f49fe94af402abe465e8415b16f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f3d88cb0eb54bc1a852c5f43f3806d9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1260.565704] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1260.565704] env[61995]: value = "resgroup-9" [ 1260.565704] env[61995]: _type = "ResourcePool" [ 1260.565704] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1260.566019] env[61995]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a95f6449-6f40-4323-a72e-22a6ee9bbdb5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.588499] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease: (returnval){ [ 1260.588499] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5229f18a-c7e9-6559-d865-b1288157c97e" [ 1260.588499] env[61995]: _type = "HttpNfcLease" [ 1260.588499] env[61995]: } obtained for vApp import into resource pool (val){ [ 1260.588499] env[61995]: value = "resgroup-9" [ 1260.588499] env[61995]: _type = "ResourcePool" [ 1260.588499] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1260.589129] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the lease: (returnval){ [ 1260.589129] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5229f18a-c7e9-6559-d865-b1288157c97e" [ 1260.589129] env[61995]: _type = "HttpNfcLease" [ 1260.589129] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1260.594876] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1260.594876] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5229f18a-c7e9-6559-d865-b1288157c97e" [ 1260.594876] env[61995]: _type = "HttpNfcLease" [ 1260.594876] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1260.731756] env[61995]: DEBUG nova.network.neutron [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updated VIF entry in instance network info cache for port 47ddce51-cfec-49c4-ba3b-b34c86140652. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1260.732169] env[61995]: DEBUG nova.network.neutron [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [{"id": "47ddce51-cfec-49c4-ba3b-b34c86140652", "address": "fa:16:3e:05:74:cc", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47ddce51-cf", "ovs_interfaceid": "47ddce51-cfec-49c4-ba3b-b34c86140652", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1260.878522] env[61995]: DEBUG nova.network.neutron [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Successfully created port: 13c49171-892c-401c-a8ea-d256ef856e2d {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1260.989185] env[61995]: DEBUG nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1260.994788] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "refresh_cache-87efc871-d14b-4b9f-98c0-87474651cbff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1260.995385] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquired lock "refresh_cache-87efc871-d14b-4b9f-98c0-87474651cbff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1260.995385] env[61995]: DEBUG nova.network.neutron [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1261.062838] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092fe866-71f2-4bdc-acdc-ecbb3f7157ac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.070630] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c6f8a0-fa80-478f-b9d6-0129f55c5001 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.102357] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981fa786-d349-41b6-bba9-8693470f7f3f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.109438] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1261.109438] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5229f18a-c7e9-6559-d865-b1288157c97e" [ 1261.109438] env[61995]: _type = "HttpNfcLease" [ 1261.109438] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1261.115782] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1261.115782] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5229f18a-c7e9-6559-d865-b1288157c97e" [ 1261.115782] env[61995]: _type = "HttpNfcLease" [ 1261.115782] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1261.115782] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b347735-7890-493e-96ef-2e22b37ac547 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.117078] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd4bd19-7643-49b1-982b-e58a5c270b9d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.124883] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52317e80-6439-a4e0-e72b-ceecffd9329c/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1261.125086] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52317e80-6439-a4e0-e72b-ceecffd9329c/disk-0.vmdk. {{(pid=61995) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1261.133716] env[61995]: DEBUG nova.compute.provider_tree [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1261.197529] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1c336d1a-eecb-4290-b182-b9937c946d5d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.235527] env[61995]: DEBUG oslo_concurrency.lockutils [req-eb070034-9e75-48fa-b76f-8ba1cba380d8 req-eb5fa154-9ac8-442e-abbe-aafadd4829ac service nova] Releasing lock "refresh_cache-24853688-b972-48e6-be7b-7f0cf634c60d" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1261.537043] env[61995]: DEBUG nova.network.neutron [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1261.636831] env[61995]: DEBUG nova.scheduler.client.report [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1261.687409] env[61995]: DEBUG nova.network.neutron [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Updating instance_info_cache with network_info: [{"id": "82531ec7-7339-4d0d-9535-099ebf8ee9b7", "address": "fa:16:3e:6e:9e:1e", "network": {"id": "80cb2c91-5cc2-4032-b955-c570bbd0ebd8", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1685649435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c706e3ad98f7431b891f4df1fffc7745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82531ec7-73", "ovs_interfaceid": "82531ec7-7339-4d0d-9535-099ebf8ee9b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1261.758456] env[61995]: DEBUG nova.compute.manager [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Received event network-changed-82531ec7-7339-4d0d-9535-099ebf8ee9b7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1261.758809] env[61995]: DEBUG nova.compute.manager [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Refreshing instance network info cache due to event network-changed-82531ec7-7339-4d0d-9535-099ebf8ee9b7. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1261.759083] env[61995]: DEBUG oslo_concurrency.lockutils [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] Acquiring lock "refresh_cache-87efc871-d14b-4b9f-98c0-87474651cbff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1262.004871] env[61995]: DEBUG nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1262.030476] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1262.030731] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1262.030893] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1262.031093] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1262.031285] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1262.031473] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1262.031690] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1262.031854] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1262.032036] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1262.032209] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1262.032387] env[61995]: DEBUG nova.virt.hardware [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1262.033382] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a90c9b3-3f5a-4331-a6d2-3fdf158b5065 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.043511] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8684b060-2eec-4fb3-98a9-7fa88b531217 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.141710] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.142279] env[61995]: DEBUG nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1262.190629] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Releasing lock "refresh_cache-87efc871-d14b-4b9f-98c0-87474651cbff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1262.191047] env[61995]: DEBUG nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Instance network_info: |[{"id": "82531ec7-7339-4d0d-9535-099ebf8ee9b7", "address": "fa:16:3e:6e:9e:1e", "network": {"id": "80cb2c91-5cc2-4032-b955-c570bbd0ebd8", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1685649435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c706e3ad98f7431b891f4df1fffc7745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82531ec7-73", "ovs_interfaceid": "82531ec7-7339-4d0d-9535-099ebf8ee9b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1262.191384] env[61995]: DEBUG oslo_concurrency.lockutils [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] Acquired lock "refresh_cache-87efc871-d14b-4b9f-98c0-87474651cbff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.191615] env[61995]: DEBUG nova.network.neutron [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Refreshing network info cache for port 82531ec7-7339-4d0d-9535-099ebf8ee9b7 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1262.193019] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:9e:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f880ac2e-d532-4f54-87bb-998a8d1bca78', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82531ec7-7339-4d0d-9535-099ebf8ee9b7', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1262.202679] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Creating folder: Project (c706e3ad98f7431b891f4df1fffc7745). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1262.207617] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08a0e8c9-8a88-4269-b933-d4ef06f1879c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.220050] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Created folder: Project (c706e3ad98f7431b891f4df1fffc7745) in parent group-v185203. [ 1262.220327] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Creating folder: Instances. Parent ref: group-v185460. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1262.220599] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3cc00f2-5081-4253-a05b-6948230df7c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.230937] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Created folder: Instances in parent group-v185460. [ 1262.285992] env[61995]: DEBUG oslo.service.loopingcall [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1262.285992] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e6afbe3-ea63-47a4-82e3-34d724b48ada {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1262.285992] env[61995]: value = "task-795541" [ 1262.285992] env[61995]: _type = "Task" [ 1262.285992] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Completed reading data from the image iterator. {{(pid=61995) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52317e80-6439-a4e0-e72b-ceecffd9329c/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7203a264-37f9-4ede-84fd-99b9e13fc028 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795541, 'name': CreateVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52317e80-6439-a4e0-e72b-ceecffd9329c/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52317e80-6439-a4e0-e72b-ceecffd9329c/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1262.285992] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-6b47780f-beaf-4170-bf11-cc01ef0d8dae {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.348555] env[61995]: DEBUG nova.compute.manager [req-5f9e4295-2602-421f-9cba-38d1e184cd7b req-bcb9ee52-79ad-4aa1-9a6d-92650e479431 service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Received event network-vif-plugged-13c49171-892c-401c-a8ea-d256ef856e2d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1262.348750] env[61995]: DEBUG oslo_concurrency.lockutils [req-5f9e4295-2602-421f-9cba-38d1e184cd7b req-bcb9ee52-79ad-4aa1-9a6d-92650e479431 service nova] Acquiring lock "47db0510-411f-415c-99a4-63ed1bcbcbc9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.349080] env[61995]: DEBUG oslo_concurrency.lockutils [req-5f9e4295-2602-421f-9cba-38d1e184cd7b req-bcb9ee52-79ad-4aa1-9a6d-92650e479431 service nova] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.349333] env[61995]: DEBUG oslo_concurrency.lockutils [req-5f9e4295-2602-421f-9cba-38d1e184cd7b req-bcb9ee52-79ad-4aa1-9a6d-92650e479431 service nova] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.349559] env[61995]: DEBUG nova.compute.manager [req-5f9e4295-2602-421f-9cba-38d1e184cd7b req-bcb9ee52-79ad-4aa1-9a6d-92650e479431 service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] No waiting events found dispatching network-vif-plugged-13c49171-892c-401c-a8ea-d256ef856e2d {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1262.349780] env[61995]: WARNING nova.compute.manager [req-5f9e4295-2602-421f-9cba-38d1e184cd7b req-bcb9ee52-79ad-4aa1-9a6d-92650e479431 service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Received unexpected event network-vif-plugged-13c49171-892c-401c-a8ea-d256ef856e2d for instance with vm_state building and task_state spawning. [ 1262.481312] env[61995]: DEBUG oslo_vmware.rw_handles [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52317e80-6439-a4e0-e72b-ceecffd9329c/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1262.481593] env[61995]: INFO nova.virt.vmwareapi.images [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Downloaded image file data 6a74f8d6-3b45-497c-92f2-1c2a3019f865 [ 1262.483485] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f894019-1fb1-4fc2-9b0e-95d323016233 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.499496] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0f6f27a-3230-4283-a219-ab95a55aac6d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.515588] env[61995]: DEBUG nova.network.neutron [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Successfully updated port: 13c49171-892c-401c-a8ea-d256ef856e2d {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1262.526280] env[61995]: INFO nova.virt.vmwareapi.images [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] The imported VM was unregistered [ 1262.528627] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Caching image {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1262.528888] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating directory with path [datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1262.529830] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c95bc3b-3f82-4888-ac12-bb0fce99bfc9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.543514] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created directory with path [datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865 {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1262.543790] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d/OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d.vmdk to [datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865/6a74f8d6-3b45-497c-92f2-1c2a3019f865.vmdk. {{(pid=61995) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1262.544093] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c58593d5-5003-4e2b-a848-85ef91d9e4c4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.554114] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1262.554114] env[61995]: value = "task-795543" [ 1262.554114] env[61995]: _type = "Task" [ 1262.554114] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.562954] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795543, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.649236] env[61995]: DEBUG nova.compute.utils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1262.651052] env[61995]: DEBUG nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1262.651052] env[61995]: DEBUG nova.network.neutron [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1262.690500] env[61995]: DEBUG nova.policy [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '478b77770ba340f9930a2571a4a04413', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '366f33a0d0a44b12a98bc8a94ffe7e4a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1262.770010] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795541, 'name': CreateVM_Task, 'duration_secs': 0.38336} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.772431] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1262.773139] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1262.773317] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.773657] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1262.774265] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2172d822-02b1-4568-a526-ec71e155c591 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.779665] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1262.779665] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]5205b8ab-fe89-113d-ea08-3d151f851952" [ 1262.779665] env[61995]: _type = "Task" [ 1262.779665] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.788380] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5205b8ab-fe89-113d-ea08-3d151f851952, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.945253] env[61995]: DEBUG nova.network.neutron [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Updated VIF entry in instance network info cache for port 82531ec7-7339-4d0d-9535-099ebf8ee9b7. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1262.945697] env[61995]: DEBUG nova.network.neutron [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Updating instance_info_cache with network_info: [{"id": "82531ec7-7339-4d0d-9535-099ebf8ee9b7", "address": "fa:16:3e:6e:9e:1e", "network": {"id": "80cb2c91-5cc2-4032-b955-c570bbd0ebd8", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1685649435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c706e3ad98f7431b891f4df1fffc7745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82531ec7-73", "ovs_interfaceid": "82531ec7-7339-4d0d-9535-099ebf8ee9b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.959697] env[61995]: DEBUG nova.network.neutron [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Successfully created port: be5b1c30-9807-4ce6-b41b-6e571c083675 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1263.017994] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-47db0510-411f-415c-99a4-63ed1bcbcbc9" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.018208] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-47db0510-411f-415c-99a4-63ed1bcbcbc9" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.018378] env[61995]: DEBUG nova.network.neutron [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1263.066577] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795543, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.153956] env[61995]: DEBUG nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1263.296634] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]5205b8ab-fe89-113d-ea08-3d151f851952, 'name': SearchDatastore_Task, 'duration_secs': 0.057838} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.297012] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1263.297299] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1263.297594] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.297780] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.298009] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1263.298337] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea2d6e01-5c11-4875-83c5-6a7553f970e8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.315628] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1263.315896] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1263.316763] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1847b5af-bdab-4c95-a0dc-477526583dc4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.324891] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1263.324891] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520fc0cb-a3f3-2d1a-cfca-fc8257e7db64" [ 1263.324891] env[61995]: _type = "Task" [ 1263.324891] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.336549] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520fc0cb-a3f3-2d1a-cfca-fc8257e7db64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.449261] env[61995]: DEBUG oslo_concurrency.lockutils [req-9bf1f141-c4f8-4554-9e0f-bdc27e3c2337 req-685638bc-a993-426f-801f-2e541229dc6b service nova] Releasing lock "refresh_cache-87efc871-d14b-4b9f-98c0-87474651cbff" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1263.553597] env[61995]: DEBUG nova.network.neutron [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1263.566313] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795543, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.755345] env[61995]: DEBUG nova.network.neutron [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Updating instance_info_cache with network_info: [{"id": "13c49171-892c-401c-a8ea-d256ef856e2d", "address": "fa:16:3e:ce:e6:30", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13c49171-89", "ovs_interfaceid": "13c49171-892c-401c-a8ea-d256ef856e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.837274] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520fc0cb-a3f3-2d1a-cfca-fc8257e7db64, 'name': SearchDatastore_Task, 'duration_secs': 0.070838} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.837916] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e139c862-eaf7-4d20-b955-0ba761c473fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.844161] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1263.844161] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ff1f2d-5731-2993-d5f9-3bcee22b06c1" [ 1263.844161] env[61995]: _type = "Task" [ 1263.844161] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.854980] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ff1f2d-5731-2993-d5f9-3bcee22b06c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.067524] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795543, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.163656] env[61995]: DEBUG nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1264.191613] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1264.191924] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1264.192101] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1264.192290] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1264.192440] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1264.192593] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1264.192805] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1264.192998] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1264.193165] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1264.193335] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1264.193509] env[61995]: DEBUG nova.virt.hardware [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1264.194407] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089d87f1-e087-4539-95d3-23aba11ea142 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.203772] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e258c85d-3068-45c4-a6a4-2db895f41d4b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.257811] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-47db0510-411f-415c-99a4-63ed1bcbcbc9" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1264.258237] env[61995]: DEBUG nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Instance network_info: |[{"id": "13c49171-892c-401c-a8ea-d256ef856e2d", "address": "fa:16:3e:ce:e6:30", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13c49171-89", "ovs_interfaceid": "13c49171-892c-401c-a8ea-d256ef856e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1264.258701] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:e6:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aef08290-001a-4ae8-aff0-1889e2211389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13c49171-892c-401c-a8ea-d256ef856e2d', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1264.266217] env[61995]: DEBUG oslo.service.loopingcall [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1264.266481] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1264.266719] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c9dae60-6783-4328-bfd6-d70287e11a35 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.288030] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1264.288030] env[61995]: value = "task-795544" [ 1264.288030] env[61995]: _type = "Task" [ 1264.288030] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.299115] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795544, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.357425] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ff1f2d-5731-2993-d5f9-3bcee22b06c1, 'name': SearchDatastore_Task, 'duration_secs': 0.069332} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.357731] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1264.358008] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 87efc871-d14b-4b9f-98c0-87474651cbff/87efc871-d14b-4b9f-98c0-87474651cbff.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1264.358326] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cde01024-b991-4f0b-9224-0b2fba317083 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.366800] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1264.366800] env[61995]: value = "task-795545" [ 1264.366800] env[61995]: _type = "Task" [ 1264.366800] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.376946] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.380185] env[61995]: DEBUG nova.compute.manager [req-62ec7667-f1a9-40f3-a415-56e837addf92 req-15f8855c-d6e5-4f2c-a13a-230ba067c7e0 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Received event network-vif-plugged-be5b1c30-9807-4ce6-b41b-6e571c083675 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1264.380449] env[61995]: DEBUG oslo_concurrency.lockutils [req-62ec7667-f1a9-40f3-a415-56e837addf92 req-15f8855c-d6e5-4f2c-a13a-230ba067c7e0 service nova] Acquiring lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.380683] env[61995]: DEBUG oslo_concurrency.lockutils [req-62ec7667-f1a9-40f3-a415-56e837addf92 req-15f8855c-d6e5-4f2c-a13a-230ba067c7e0 service nova] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.380862] env[61995]: DEBUG oslo_concurrency.lockutils [req-62ec7667-f1a9-40f3-a415-56e837addf92 req-15f8855c-d6e5-4f2c-a13a-230ba067c7e0 service nova] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.381234] env[61995]: DEBUG nova.compute.manager [req-62ec7667-f1a9-40f3-a415-56e837addf92 req-15f8855c-d6e5-4f2c-a13a-230ba067c7e0 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] No waiting events found dispatching network-vif-plugged-be5b1c30-9807-4ce6-b41b-6e571c083675 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1264.381455] env[61995]: WARNING nova.compute.manager [req-62ec7667-f1a9-40f3-a415-56e837addf92 req-15f8855c-d6e5-4f2c-a13a-230ba067c7e0 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Received unexpected event network-vif-plugged-be5b1c30-9807-4ce6-b41b-6e571c083675 for instance with vm_state building and task_state spawning. [ 1264.382580] env[61995]: DEBUG nova.compute.manager [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Received event network-changed-13c49171-892c-401c-a8ea-d256ef856e2d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1264.382792] env[61995]: DEBUG nova.compute.manager [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Refreshing instance network info cache due to event network-changed-13c49171-892c-401c-a8ea-d256ef856e2d. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1264.383024] env[61995]: DEBUG oslo_concurrency.lockutils [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] Acquiring lock "refresh_cache-47db0510-411f-415c-99a4-63ed1bcbcbc9" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.383220] env[61995]: DEBUG oslo_concurrency.lockutils [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] Acquired lock "refresh_cache-47db0510-411f-415c-99a4-63ed1bcbcbc9" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.383409] env[61995]: DEBUG nova.network.neutron [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Refreshing network info cache for port 13c49171-892c-401c-a8ea-d256ef856e2d {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1264.441674] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1264.442089] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1264.567835] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795543, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.798093] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795544, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.876241] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.911810] env[61995]: DEBUG nova.network.neutron [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Successfully updated port: be5b1c30-9807-4ce6-b41b-6e571c083675 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1264.948532] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1264.948532] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1265.070746] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795543, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.319057} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.071277] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d/OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d.vmdk to [datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865/6a74f8d6-3b45-497c-92f2-1c2a3019f865.vmdk. [ 1265.071277] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Cleaning up location [datastore2] OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1265.071456] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_15991b06-cf6f-4f5b-816a-b9e3d796d14d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1265.071660] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc966e57-2452-44c6-9a4a-41860e6b323c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.079580] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1265.079580] env[61995]: value = "task-795546" [ 1265.079580] env[61995]: _type = "Task" [ 1265.079580] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.089780] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795546, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.098725] env[61995]: DEBUG nova.network.neutron [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Updated VIF entry in instance network info cache for port 13c49171-892c-401c-a8ea-d256ef856e2d. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1265.099188] env[61995]: DEBUG nova.network.neutron [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Updating instance_info_cache with network_info: [{"id": "13c49171-892c-401c-a8ea-d256ef856e2d", "address": "fa:16:3e:ce:e6:30", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13c49171-89", "ovs_interfaceid": "13c49171-892c-401c-a8ea-d256ef856e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.298664] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795544, 'name': CreateVM_Task, 'duration_secs': 0.852137} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.298848] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1265.299480] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1265.299650] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.299976] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1265.300241] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c871d69f-cd67-4d2c-8c29-d0d3bfa735e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.304303] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1265.304303] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52dea56c-ac78-73a7-08f8-ca954974c310" [ 1265.304303] env[61995]: _type = "Task" [ 1265.304303] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.311374] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52dea56c-ac78-73a7-08f8-ca954974c310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.376857] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.912432} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.377134] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 87efc871-d14b-4b9f-98c0-87474651cbff/87efc871-d14b-4b9f-98c0-87474651cbff.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1265.377351] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1265.377600] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed78500b-1f59-4561-b790-16331c1c7d91 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.384138] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1265.384138] env[61995]: value = "task-795547" [ 1265.384138] env[61995]: _type = "Task" [ 1265.384138] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.391047] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.418443] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "refresh_cache-95af4a7a-bfe9-4bad-bb61-1df3736f81dc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1265.418545] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquired lock "refresh_cache-95af4a7a-bfe9-4bad-bb61-1df3736f81dc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.418669] env[61995]: DEBUG nova.network.neutron [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1265.453701] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Didn't find any instances for network info cache update. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1265.453908] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.454095] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.454252] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.454405] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.454548] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.454693] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.454822] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1265.454963] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.589818] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795546, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180122} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.590027] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1265.590214] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865/6a74f8d6-3b45-497c-92f2-1c2a3019f865.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.590530] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865/6a74f8d6-3b45-497c-92f2-1c2a3019f865.vmdk to [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1265.590793] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8576bc06-cf34-4dfd-8f32-16fc61dbcfcb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.597935] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1265.597935] env[61995]: value = "task-795548" [ 1265.597935] env[61995]: _type = "Task" [ 1265.597935] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.602455] env[61995]: DEBUG oslo_concurrency.lockutils [req-cfc8de9c-9f03-4cdf-9f97-9c256a0fed3c req-aed28949-3032-47d3-82d5-ff897390704c service nova] Releasing lock "refresh_cache-47db0510-411f-415c-99a4-63ed1bcbcbc9" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.607319] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.819480] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52dea56c-ac78-73a7-08f8-ca954974c310, 'name': SearchDatastore_Task, 'duration_secs': 0.008758} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.819987] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.820405] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1265.820818] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1265.821118] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.821461] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1265.821879] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3b904d4-8add-4592-9087-b0f52a2f1c80 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.832746] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1265.832969] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1265.833713] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5404002b-2bda-49c6-86c0-101e3c8f2e96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.839606] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1265.839606] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52118c24-9f39-3cc3-6ebe-f1212f8b3c08" [ 1265.839606] env[61995]: _type = "Task" [ 1265.839606] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.848034] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52118c24-9f39-3cc3-6ebe-f1212f8b3c08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.894996] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061036} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.895174] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1265.895905] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f99ca43-bbb8-461e-b5f8-112bce246ced {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.918580] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 87efc871-d14b-4b9f-98c0-87474651cbff/87efc871-d14b-4b9f-98c0-87474651cbff.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1265.918897] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2a3fb1c-bb48-4354-9b4b-c8924d19b3fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.942224] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1265.942224] env[61995]: value = "task-795549" [ 1265.942224] env[61995]: _type = "Task" [ 1265.942224] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.950177] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795549, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.958228] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.958456] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.958617] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.958767] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1265.959594] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416729c9-44a7-4554-a9b3-58b426d02c4d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.967272] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382ee29d-5743-4589-b3c2-0ada8b4cd3f1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.981998] env[61995]: DEBUG nova.network.neutron [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1265.984475] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51a1d17-2644-452d-a4c2-1b6e0a1b1aff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.991476] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7dc51e0-0dfd-46b8-91df-2324332cda22 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.021940] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181261MB free_disk=66GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1266.022116] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.022285] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.108049] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.143151] env[61995]: DEBUG nova.network.neutron [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Updating instance_info_cache with network_info: [{"id": "be5b1c30-9807-4ce6-b41b-6e571c083675", "address": "fa:16:3e:34:b1:d6", "network": {"id": "4cd54505-33a1-4912-9042-a67bd442a63b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1860577328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "366f33a0d0a44b12a98bc8a94ffe7e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe5b1c30-98", "ovs_interfaceid": "be5b1c30-9807-4ce6-b41b-6e571c083675", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.350083] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52118c24-9f39-3cc3-6ebe-f1212f8b3c08, 'name': SearchDatastore_Task, 'duration_secs': 0.012117} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.350892] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a8ce1c9-d1e5-47c2-a580-aafc284dd414 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.356138] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1266.356138] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]523ff67a-608d-aa2c-8e97-eb0bb3211d07" [ 1266.356138] env[61995]: _type = "Task" [ 1266.356138] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.363684] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523ff67a-608d-aa2c-8e97-eb0bb3211d07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.414264] env[61995]: DEBUG nova.compute.manager [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Received event network-changed-be5b1c30-9807-4ce6-b41b-6e571c083675 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1266.414470] env[61995]: DEBUG nova.compute.manager [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Refreshing instance network info cache due to event network-changed-be5b1c30-9807-4ce6-b41b-6e571c083675. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1266.414661] env[61995]: DEBUG oslo_concurrency.lockutils [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] Acquiring lock "refresh_cache-95af4a7a-bfe9-4bad-bb61-1df3736f81dc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1266.453025] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795549, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.608903] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.645656] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Releasing lock "refresh_cache-95af4a7a-bfe9-4bad-bb61-1df3736f81dc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1266.646012] env[61995]: DEBUG nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Instance network_info: |[{"id": "be5b1c30-9807-4ce6-b41b-6e571c083675", "address": "fa:16:3e:34:b1:d6", "network": {"id": "4cd54505-33a1-4912-9042-a67bd442a63b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1860577328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "366f33a0d0a44b12a98bc8a94ffe7e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe5b1c30-98", "ovs_interfaceid": "be5b1c30-9807-4ce6-b41b-6e571c083675", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1266.646394] env[61995]: DEBUG oslo_concurrency.lockutils [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] Acquired lock "refresh_cache-95af4a7a-bfe9-4bad-bb61-1df3736f81dc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1266.646587] env[61995]: DEBUG nova.network.neutron [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Refreshing network info cache for port be5b1c30-9807-4ce6-b41b-6e571c083675 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1266.647992] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:b1:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31a7f15-a808-4199-9071-31fd05e316ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be5b1c30-9807-4ce6-b41b-6e571c083675', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1266.655575] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Creating folder: Project (366f33a0d0a44b12a98bc8a94ffe7e4a). Parent ref: group-v185203. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1266.659037] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5657a289-e31d-4dd6-b31d-9e24ba51b53a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.671995] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Created folder: Project (366f33a0d0a44b12a98bc8a94ffe7e4a) in parent group-v185203. [ 1266.672243] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Creating folder: Instances. Parent ref: group-v185464. {{(pid=61995) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1266.672507] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa3e2b96-46db-4a59-b830-100b5cafe6c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.683146] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Created folder: Instances in parent group-v185464. [ 1266.683419] env[61995]: DEBUG oslo.service.loopingcall [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1266.683619] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1266.683829] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d30ae191-508c-4102-9132-8125e288e213 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.708354] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1266.708354] env[61995]: value = "task-795552" [ 1266.708354] env[61995]: _type = "Task" [ 1266.708354] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.716999] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795552, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.868217] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]523ff67a-608d-aa2c-8e97-eb0bb3211d07, 'name': SearchDatastore_Task, 'duration_secs': 0.039759} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.868521] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1266.868770] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 47db0510-411f-415c-99a4-63ed1bcbcbc9/47db0510-411f-415c-99a4-63ed1bcbcbc9.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1266.869057] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7918ed28-3ec7-4e81-8e3f-fb16cba93675 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.878399] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1266.878399] env[61995]: value = "task-795553" [ 1266.878399] env[61995]: _type = "Task" [ 1266.878399] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.883509] env[61995]: DEBUG nova.network.neutron [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Updated VIF entry in instance network info cache for port be5b1c30-9807-4ce6-b41b-6e571c083675. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1266.883897] env[61995]: DEBUG nova.network.neutron [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Updating instance_info_cache with network_info: [{"id": "be5b1c30-9807-4ce6-b41b-6e571c083675", "address": "fa:16:3e:34:b1:d6", "network": {"id": "4cd54505-33a1-4912-9042-a67bd442a63b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1860577328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "366f33a0d0a44b12a98bc8a94ffe7e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe5b1c30-98", "ovs_interfaceid": "be5b1c30-9807-4ce6-b41b-6e571c083675", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.888278] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.955932] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795549, 'name': ReconfigVM_Task, 'duration_secs': 0.707702} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.956347] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 87efc871-d14b-4b9f-98c0-87474651cbff/87efc871-d14b-4b9f-98c0-87474651cbff.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1266.957015] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea5fdd15-e00e-42cd-8486-76c78b463a20 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.966123] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1266.966123] env[61995]: value = "task-795554" [ 1266.966123] env[61995]: _type = "Task" [ 1266.966123] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.977816] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795554, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.052221] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 24853688-b972-48e6-be7b-7f0cf634c60d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.052383] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 87efc871-d14b-4b9f-98c0-87474651cbff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.052505] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 47db0510-411f-415c-99a4-63ed1bcbcbc9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.052620] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 95af4a7a-bfe9-4bad-bb61-1df3736f81dc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.052826] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1267.052933] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1267.112468] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.127077] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d826abd3-b8f9-45e2-8822-f9dbc8fc2e4e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.137049] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e63590d-64df-436a-b9a6-7bfe2e619567 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.169013] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6d3ee9-3cc3-4f4b-b0d4-c22bc91f4e70 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.179129] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b72d43-585b-47ef-b6b4-07c1fe93bd89 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.193530] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1267.220136] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795552, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.390899] env[61995]: DEBUG oslo_concurrency.lockutils [req-27cd2ed5-ed09-479b-838f-904e655d59b8 req-cc86a17c-6229-4764-8787-40323845be15 service nova] Releasing lock "refresh_cache-95af4a7a-bfe9-4bad-bb61-1df3736f81dc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1267.391344] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.478287] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795554, 'name': Rename_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.612557] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.696966] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1267.725222] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795552, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.892114] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.977794] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795554, 'name': Rename_Task, 'duration_secs': 0.793888} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.978113] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1267.978444] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e34d5b69-7395-436a-a0c3-1e2f853f0aa2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.987467] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1267.987467] env[61995]: value = "task-795555" [ 1267.987467] env[61995]: _type = "Task" [ 1267.987467] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.997802] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.113975] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.202713] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1268.202713] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.180s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.222694] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795552, 'name': CreateVM_Task} progress is 25%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.394321] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795553, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.499753] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795555, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.612631] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.722199] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795552, 'name': CreateVM_Task, 'duration_secs': 1.802526} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.722428] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1268.723107] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1268.723293] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1268.723649] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1268.723927] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ec721b6-633a-4791-a71d-7d92d157bc9e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.729943] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1268.729943] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52a4ff9f-a935-758d-5d40-850577cfd9c7" [ 1268.729943] env[61995]: _type = "Task" [ 1268.729943] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.740520] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a4ff9f-a935-758d-5d40-850577cfd9c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.891733] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795553, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.998645] env[61995]: DEBUG oslo_vmware.api [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795555, 'name': PowerOnVM_Task, 'duration_secs': 0.53383} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.998994] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1268.999129] env[61995]: INFO nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Took 9.49 seconds to spawn the instance on the hypervisor. [ 1268.999319] env[61995]: DEBUG nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1269.000126] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db45e3f8-ad85-4917-bd22-483922687855 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.112486] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795548, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.380034} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.112691] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6a74f8d6-3b45-497c-92f2-1c2a3019f865/6a74f8d6-3b45-497c-92f2-1c2a3019f865.vmdk to [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1269.113433] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67382feb-921a-47ca-9de9-1a7fc3306626 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.136390] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1269.136620] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-011fc9d6-110c-4dd0-bcb1-e95366631fc5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.155041] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1269.155041] env[61995]: value = "task-795556" [ 1269.155041] env[61995]: _type = "Task" [ 1269.155041] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.162293] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.240986] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52a4ff9f-a935-758d-5d40-850577cfd9c7, 'name': SearchDatastore_Task, 'duration_secs': 0.121017} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.241360] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1269.241655] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1269.241955] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1269.242199] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.242446] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1269.242762] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7938110b-efa2-4bcb-88f5-6084a9ffe202 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.250689] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1269.250879] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1269.251558] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0802c7f5-574c-4ab2-857d-72e44e3e33b8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.256126] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1269.256126] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]522f4a35-9236-bcb3-9e6b-fba6117ed0e5" [ 1269.256126] env[61995]: _type = "Task" [ 1269.256126] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.263423] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522f4a35-9236-bcb3-9e6b-fba6117ed0e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.391115] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795553, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.114987} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.391397] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 47db0510-411f-415c-99a4-63ed1bcbcbc9/47db0510-411f-415c-99a4-63ed1bcbcbc9.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1269.391611] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1269.391932] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-034cd0d4-0e18-4422-948a-4701ba759585 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.399122] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1269.399122] env[61995]: value = "task-795557" [ 1269.399122] env[61995]: _type = "Task" [ 1269.399122] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.406860] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.515629] env[61995]: INFO nova.compute.manager [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Took 17.04 seconds to build instance. [ 1269.665588] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795556, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.766322] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]522f4a35-9236-bcb3-9e6b-fba6117ed0e5, 'name': SearchDatastore_Task, 'duration_secs': 0.007695} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.767118] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62a464bb-a3dd-4c59-ba83-7bf58d669e73 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.771985] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1269.771985] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52812103-1eb5-aa74-0c0b-d75482648312" [ 1269.771985] env[61995]: _type = "Task" [ 1269.771985] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.779503] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52812103-1eb5-aa74-0c0b-d75482648312, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.908708] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07096} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.908980] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1269.909725] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a516885-39be-45a3-9ad3-7adf320a4dee {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.930992] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 47db0510-411f-415c-99a4-63ed1bcbcbc9/47db0510-411f-415c-99a4-63ed1bcbcbc9.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1269.931310] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee53bb61-7ab0-4db3-905f-ea649cd06d92 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.950717] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1269.950717] env[61995]: value = "task-795558" [ 1269.950717] env[61995]: _type = "Task" [ 1269.950717] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.958944] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795558, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.017660] env[61995]: DEBUG oslo_concurrency.lockutils [None req-cc2071cd-ffbd-4b2b-bff9-86480c55c8f1 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "87efc871-d14b-4b9f-98c0-87474651cbff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.555s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.167244] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795556, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.283520] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52812103-1eb5-aa74-0c0b-d75482648312, 'name': SearchDatastore_Task, 'duration_secs': 0.008528} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.283823] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1270.284512] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 95af4a7a-bfe9-4bad-bb61-1df3736f81dc/95af4a7a-bfe9-4bad-bb61-1df3736f81dc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1270.284512] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf19b589-a213-402b-9bbf-e1ddc346f282 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.291958] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1270.291958] env[61995]: value = "task-795559" [ 1270.291958] env[61995]: _type = "Task" [ 1270.291958] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.299997] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.461806] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795558, 'name': ReconfigVM_Task, 'duration_secs': 0.297557} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.462135] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 47db0510-411f-415c-99a4-63ed1bcbcbc9/47db0510-411f-415c-99a4-63ed1bcbcbc9.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1270.462798] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc5584ac-00ff-4d9b-99bc-29bc78cafe24 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.470269] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1270.470269] env[61995]: value = "task-795560" [ 1270.470269] env[61995]: _type = "Task" [ 1270.470269] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.480855] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795560, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.667926] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795556, 'name': ReconfigVM_Task, 'duration_secs': 1.093655} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.668338] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d/24853688-b972-48e6-be7b-7f0cf634c60d.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1270.669463] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_format': None, 'device_type': 'disk', 'encryption_secret_uuid': None, 'device_name': '/dev/sda', 'encryption_options': None, 'guest_format': None, 'encrypted': False, 'boot_index': 0, 'size': 0, 'disk_bus': None, 'image_id': 'c1739132-aff3-4644-9522-fc02cf765bc8'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185457', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'name': 'volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '24853688-b972-48e6-be7b-7f0cf634c60d', 'attached_at': '', 'detached_at': '', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'serial': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed'}, 'attachment_id': 'e696a5fa-c189-42ef-8fbb-91cc54576083', 'guest_format': None, 'boot_index': None, 'disk_bus': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=61995) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1270.669716] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Volume attach. Driver type: vmdk {{(pid=61995) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1270.669925] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185457', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'name': 'volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '24853688-b972-48e6-be7b-7f0cf634c60d', 'attached_at': '', 'detached_at': '', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'serial': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1270.670731] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1dca2eb-553b-4f8e-a895-1be385cd66fa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.688464] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8ebaa8-2724-4fd1-88cb-bdaca605c87d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.713158] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed/volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1270.713506] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7029913-5979-4e26-b24f-f56afcd6f9c0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.731379] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1270.731379] env[61995]: value = "task-795561" [ 1270.731379] env[61995]: _type = "Task" [ 1270.731379] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.739672] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795561, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.800448] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795559, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421647} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.800714] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 95af4a7a-bfe9-4bad-bb61-1df3736f81dc/95af4a7a-bfe9-4bad-bb61-1df3736f81dc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1270.800926] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1270.801193] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-003821b0-e78b-4330-9e66-56a65c8cd188 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.807467] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1270.807467] env[61995]: value = "task-795562" [ 1270.807467] env[61995]: _type = "Task" [ 1270.807467] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.814920] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795562, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.854450] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "87efc871-d14b-4b9f-98c0-87474651cbff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1270.854693] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "87efc871-d14b-4b9f-98c0-87474651cbff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.854908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "87efc871-d14b-4b9f-98c0-87474651cbff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1270.855112] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "87efc871-d14b-4b9f-98c0-87474651cbff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.855293] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "87efc871-d14b-4b9f-98c0-87474651cbff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.857416] env[61995]: INFO nova.compute.manager [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Terminating instance [ 1270.859304] env[61995]: DEBUG nova.compute.manager [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1270.859498] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1270.860258] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e495b5-a71d-4c86-bd02-601273b1f236 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.867386] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1270.867627] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3fc30705-dc90-4495-8c63-f9d4ef8b0ab7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.875600] env[61995]: DEBUG oslo_vmware.api [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1270.875600] env[61995]: value = "task-795563" [ 1270.875600] env[61995]: _type = "Task" [ 1270.875600] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.883179] env[61995]: DEBUG oslo_vmware.api [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795563, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.981350] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795560, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.241636] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795561, 'name': ReconfigVM_Task, 'duration_secs': 0.292022} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.243069] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed/volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1271.246928] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-747343f2-bd73-4afc-a9c9-e3d750f6633f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.261328] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1271.261328] env[61995]: value = "task-795564" [ 1271.261328] env[61995]: _type = "Task" [ 1271.261328] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.272246] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795564, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.316990] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795562, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.384921] env[61995]: DEBUG oslo_vmware.api [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795563, 'name': PowerOffVM_Task, 'duration_secs': 0.184869} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.385201] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1271.385376] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1271.385633] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7851e210-9d11-4290-8cfa-42cf2b68cd29 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.480786] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795560, 'name': Rename_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.488026] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1271.488026] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1271.488231] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Deleting the datastore file [datastore2] 87efc871-d14b-4b9f-98c0-87474651cbff {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1271.488341] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39378302-cd61-45b5-ac3d-a1ee7ac9289f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.494895] env[61995]: DEBUG oslo_vmware.api [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for the task: (returnval){ [ 1271.494895] env[61995]: value = "task-795566" [ 1271.494895] env[61995]: _type = "Task" [ 1271.494895] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.502254] env[61995]: DEBUG oslo_vmware.api [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.771255] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795564, 'name': ReconfigVM_Task, 'duration_secs': 0.143579} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.771589] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185457', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'name': 'volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '24853688-b972-48e6-be7b-7f0cf634c60d', 'attached_at': '', 'detached_at': '', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'serial': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed'} {{(pid=61995) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1271.772196] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-573ed6dd-6d3e-43a6-b1f9-2c481fac7625 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.779092] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1271.779092] env[61995]: value = "task-795567" [ 1271.779092] env[61995]: _type = "Task" [ 1271.779092] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.786859] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795567, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.817248] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795562, 'name': ExtendVirtualDisk_Task} progress is 100%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.981112] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795560, 'name': Rename_Task, 'duration_secs': 1.319247} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.981404] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1271.981644] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d8afbc5-494d-497f-9b75-60e0e62c9d6a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.987498] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1271.987498] env[61995]: value = "task-795568" [ 1271.987498] env[61995]: _type = "Task" [ 1271.987498] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.994470] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.002253] env[61995]: DEBUG oslo_vmware.api [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Task: {'id': task-795566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153369} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.002485] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1272.002672] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1272.002848] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1272.003037] env[61995]: INFO nova.compute.manager [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1272.003298] env[61995]: DEBUG oslo.service.loopingcall [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1272.003498] env[61995]: DEBUG nova.compute.manager [-] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1272.003593] env[61995]: DEBUG nova.network.neutron [-] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1272.245931] env[61995]: DEBUG nova.compute.manager [req-c00ca4cf-a34f-45e4-9718-c1899ae99427 req-8f7ad649-eb87-444a-a8bf-fb9c282cc485 service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Received event network-vif-deleted-82531ec7-7339-4d0d-9535-099ebf8ee9b7 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1272.246286] env[61995]: INFO nova.compute.manager [req-c00ca4cf-a34f-45e4-9718-c1899ae99427 req-8f7ad649-eb87-444a-a8bf-fb9c282cc485 service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Neutron deleted interface 82531ec7-7339-4d0d-9535-099ebf8ee9b7; detaching it from the instance and deleting it from the info cache [ 1272.246332] env[61995]: DEBUG nova.network.neutron [req-c00ca4cf-a34f-45e4-9718-c1899ae99427 req-8f7ad649-eb87-444a-a8bf-fb9c282cc485 service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1272.289512] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795567, 'name': Rename_Task, 'duration_secs': 0.174641} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.289758] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1272.290011] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9cfe792-20f9-4494-bfef-24ca5ebb19c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.296457] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1272.296457] env[61995]: value = "task-795569" [ 1272.296457] env[61995]: _type = "Task" [ 1272.296457] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.317780] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795569, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.323607] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795562, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.012445} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.323881] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1272.324658] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27631b02-6bcd-4431-a217-2f63e4485f37 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.350799] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] 95af4a7a-bfe9-4bad-bb61-1df3736f81dc/95af4a7a-bfe9-4bad-bb61-1df3736f81dc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1272.351291] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e2e12ba-d882-4368-86b1-5973aaf4957f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.380872] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1272.380872] env[61995]: value = "task-795570" [ 1272.380872] env[61995]: _type = "Task" [ 1272.380872] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.391664] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795570, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.496890] env[61995]: DEBUG oslo_vmware.api [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795568, 'name': PowerOnVM_Task, 'duration_secs': 0.469861} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.497178] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1272.497414] env[61995]: INFO nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Took 10.49 seconds to spawn the instance on the hypervisor. [ 1272.497592] env[61995]: DEBUG nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1272.498397] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204a4909-1a18-46a4-b53f-8f4e09af0fbe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.718910] env[61995]: DEBUG nova.network.neutron [-] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1272.749476] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0184761c-9791-46f3-94fa-dfff03a8c1f7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.758524] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2335e1-05e8-4c81-8573-bc99286eeeaa {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.783142] env[61995]: DEBUG nova.compute.manager [req-c00ca4cf-a34f-45e4-9718-c1899ae99427 req-8f7ad649-eb87-444a-a8bf-fb9c282cc485 service nova] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Detach interface failed, port_id=82531ec7-7339-4d0d-9535-099ebf8ee9b7, reason: Instance 87efc871-d14b-4b9f-98c0-87474651cbff could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1272.806095] env[61995]: DEBUG oslo_vmware.api [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795569, 'name': PowerOnVM_Task, 'duration_secs': 0.4724} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.806095] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1272.895023] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795570, 'name': ReconfigVM_Task, 'duration_secs': 0.305352} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.895023] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Reconfigured VM instance instance-00000074 to attach disk [datastore2] 95af4a7a-bfe9-4bad-bb61-1df3736f81dc/95af4a7a-bfe9-4bad-bb61-1df3736f81dc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1272.895023] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cab9322c-5c2f-4c18-9382-3671c65c62b2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.903131] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1272.903131] env[61995]: value = "task-795571" [ 1272.903131] env[61995]: _type = "Task" [ 1272.903131] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.914686] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795571, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.930393] env[61995]: DEBUG nova.compute.manager [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1272.931350] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4682ca3-d530-48d6-add3-6884e57c7895 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.016616] env[61995]: INFO nova.compute.manager [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Took 18.73 seconds to build instance. [ 1273.221290] env[61995]: INFO nova.compute.manager [-] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Took 1.22 seconds to deallocate network for instance. [ 1273.413890] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795571, 'name': Rename_Task, 'duration_secs': 0.179495} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.414450] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1273.414450] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3811f53b-b394-4df8-bb14-7df9f978b6c1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.420482] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1273.420482] env[61995]: value = "task-795572" [ 1273.420482] env[61995]: _type = "Task" [ 1273.420482] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.425236] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.428137] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.450895] env[61995]: DEBUG oslo_concurrency.lockutils [None req-281c9488-5e3d-4a49-854f-7723b93940d4 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.389s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.518769] env[61995]: DEBUG oslo_concurrency.lockutils [None req-53950ac2-f4a3-4223-b350-5880ec3a98fc tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.243s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.519061] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.094s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.519347] env[61995]: DEBUG nova.compute.manager [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1273.520123] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8870ade9-d5ee-46c9-862b-09477d982044 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.527100] env[61995]: DEBUG nova.compute.manager [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61995) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1273.527648] env[61995]: DEBUG nova.objects.instance [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'flavor' on Instance uuid 47db0510-411f-415c-99a4-63ed1bcbcbc9 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1273.728362] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.728702] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.728961] env[61995]: DEBUG nova.objects.instance [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lazy-loading 'resources' on Instance uuid 87efc871-d14b-4b9f-98c0-87474651cbff {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1273.930874] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795572, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.032792] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1274.033066] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc6249d5-4124-4cb1-9451-e01d690e06cb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.039648] env[61995]: DEBUG oslo_vmware.api [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1274.039648] env[61995]: value = "task-795573" [ 1274.039648] env[61995]: _type = "Task" [ 1274.039648] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.047480] env[61995]: DEBUG oslo_vmware.api [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.298219] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ad330a-21cb-4cd6-8dd3-25096ed803c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.305979] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6b78b0-8890-4183-ab1a-d52c77781fb7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.336829] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af235de-ed4b-4950-bc4b-6af0313a6287 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.344057] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280607a1-11ca-4d72-b4ea-5205cbd36a4f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.356847] env[61995]: DEBUG nova.compute.provider_tree [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1274.431074] env[61995]: DEBUG oslo_vmware.api [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795572, 'name': PowerOnVM_Task, 'duration_secs': 0.842785} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.431432] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1274.431551] env[61995]: INFO nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Took 10.27 seconds to spawn the instance on the hypervisor. [ 1274.431732] env[61995]: DEBUG nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1274.432500] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9155b5b8-002e-4708-9cb2-a3c9e28e5617 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.549717] env[61995]: DEBUG oslo_vmware.api [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795573, 'name': PowerOffVM_Task, 'duration_secs': 0.162433} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.549932] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1274.549987] env[61995]: DEBUG nova.compute.manager [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1274.551029] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b276881f-1281-4cbb-a8a9-6108561a92fd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.878420] env[61995]: ERROR nova.scheduler.client.report [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] [req-92da1f19-3a20-49a4-8e62-05d755c4957a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-92da1f19-3a20-49a4-8e62-05d755c4957a"}]} [ 1274.894176] env[61995]: DEBUG nova.scheduler.client.report [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1274.906794] env[61995]: DEBUG nova.scheduler.client.report [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1274.907007] env[61995]: DEBUG nova.compute.provider_tree [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1274.919532] env[61995]: DEBUG nova.scheduler.client.report [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1274.936880] env[61995]: DEBUG nova.scheduler.client.report [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1274.948110] env[61995]: INFO nova.compute.manager [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Took 18.22 seconds to build instance. [ 1275.000659] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f3fc89-ca19-4f38-a66e-c1a0945da6bb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.007750] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587930af-8374-4814-b61b-c298e95bc83a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.038185] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe741a0b-31f1-4daf-a8cc-9d7ad21bb3f4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.045129] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6112695-c162-455d-88bb-ae519bfcb37c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.058557] env[61995]: DEBUG nova.compute.provider_tree [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1275.063771] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6869e5fc-497f-46fb-b26d-0c87cde50c67 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.450389] env[61995]: DEBUG oslo_concurrency.lockutils [None req-a42ca8ca-225f-4472-b8c0-da6fba09e006 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.735s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.588452] env[61995]: DEBUG nova.scheduler.client.report [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 163 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1275.588822] env[61995]: DEBUG nova.compute.provider_tree [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 163 to 164 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1275.589012] env[61995]: DEBUG nova.compute.provider_tree [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1275.846578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1275.846578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1275.846578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "47db0510-411f-415c-99a4-63ed1bcbcbc9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1275.846578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1275.846578] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.848309] env[61995]: INFO nova.compute.manager [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Terminating instance [ 1275.850189] env[61995]: DEBUG nova.compute.manager [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1275.850380] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1275.851229] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8418ec-be95-438e-9744-ba0d8df61d2d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.858959] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1275.859203] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c547bf4-ecdf-4c62-b36e-8326611d57f3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.095599] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.367s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.113339] env[61995]: INFO nova.scheduler.client.report [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Deleted allocations for instance 87efc871-d14b-4b9f-98c0-87474651cbff [ 1276.219540] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1276.219755] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1276.219755] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleting the datastore file [datastore2] 47db0510-411f-415c-99a4-63ed1bcbcbc9 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1276.220379] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b6c91f5-bfa1-47ee-baba-d805e889531a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.228115] env[61995]: DEBUG oslo_vmware.api [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1276.228115] env[61995]: value = "task-795575" [ 1276.228115] env[61995]: _type = "Task" [ 1276.228115] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.236230] env[61995]: DEBUG oslo_vmware.api [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.458227] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.458781] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.458781] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.458931] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.459079] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.461417] env[61995]: INFO nova.compute.manager [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Terminating instance [ 1276.463182] env[61995]: DEBUG nova.compute.manager [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1276.463380] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1276.464189] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6374cf3-18f6-482c-af57-9ccd5b1079ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.471788] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1276.472021] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5390bafd-0f0b-4d76-b53a-1efdd0d9b27b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.478568] env[61995]: DEBUG oslo_vmware.api [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1276.478568] env[61995]: value = "task-795576" [ 1276.478568] env[61995]: _type = "Task" [ 1276.478568] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.490286] env[61995]: DEBUG oslo_vmware.api [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.621395] env[61995]: DEBUG oslo_concurrency.lockutils [None req-9496fc25-edf0-4661-9770-9117ca607055 tempest-ServerMetadataNegativeTestJSON-508644362 tempest-ServerMetadataNegativeTestJSON-508644362-project-member] Lock "87efc871-d14b-4b9f-98c0-87474651cbff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.766s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.737723] env[61995]: DEBUG oslo_vmware.api [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.378161} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.737976] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1276.738195] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1276.738374] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1276.738555] env[61995]: INFO nova.compute.manager [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Took 0.89 seconds to destroy the instance on the hypervisor. [ 1276.738795] env[61995]: DEBUG oslo.service.loopingcall [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1276.738978] env[61995]: DEBUG nova.compute.manager [-] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1276.739088] env[61995]: DEBUG nova.network.neutron [-] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1276.986779] env[61995]: DEBUG nova.compute.manager [req-e3593eb6-82e9-4653-99b3-af5275518a97 req-4fc0df75-1c37-4e80-9cb2-7e795cea9831 service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Received event network-vif-deleted-13c49171-892c-401c-a8ea-d256ef856e2d {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1276.987271] env[61995]: INFO nova.compute.manager [req-e3593eb6-82e9-4653-99b3-af5275518a97 req-4fc0df75-1c37-4e80-9cb2-7e795cea9831 service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Neutron deleted interface 13c49171-892c-401c-a8ea-d256ef856e2d; detaching it from the instance and deleting it from the info cache [ 1276.987271] env[61995]: DEBUG nova.network.neutron [req-e3593eb6-82e9-4653-99b3-af5275518a97 req-4fc0df75-1c37-4e80-9cb2-7e795cea9831 service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1276.991617] env[61995]: DEBUG oslo_vmware.api [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795576, 'name': PowerOffVM_Task, 'duration_secs': 0.188929} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.992231] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1276.992450] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1276.992850] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83fa7b50-12de-4bb1-bf00-f7b4faa9f2d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.131522] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1277.131784] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1277.131973] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Deleting the datastore file [datastore2] 95af4a7a-bfe9-4bad-bb61-1df3736f81dc {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1277.132252] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25916745-f1f0-45e2-b522-469d18a9aac0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.138437] env[61995]: DEBUG oslo_vmware.api [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for the task: (returnval){ [ 1277.138437] env[61995]: value = "task-795578" [ 1277.138437] env[61995]: _type = "Task" [ 1277.138437] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.146313] env[61995]: DEBUG oslo_vmware.api [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795578, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.463227] env[61995]: DEBUG nova.network.neutron [-] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1277.492986] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc5c0d81-e287-4307-a717-210e629166a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.503301] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba332c0-4c62-4b9d-908f-87a4b06908bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.527789] env[61995]: DEBUG nova.compute.manager [req-e3593eb6-82e9-4653-99b3-af5275518a97 req-4fc0df75-1c37-4e80-9cb2-7e795cea9831 service nova] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Detach interface failed, port_id=13c49171-892c-401c-a8ea-d256ef856e2d, reason: Instance 47db0510-411f-415c-99a4-63ed1bcbcbc9 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1277.648389] env[61995]: DEBUG oslo_vmware.api [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Task: {'id': task-795578, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196272} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.648674] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1277.648869] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1277.649067] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1277.649259] env[61995]: INFO nova.compute.manager [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1277.649541] env[61995]: DEBUG oslo.service.loopingcall [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1277.649794] env[61995]: DEBUG nova.compute.manager [-] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1277.649847] env[61995]: DEBUG nova.network.neutron [-] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1277.967340] env[61995]: INFO nova.compute.manager [-] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Took 1.23 seconds to deallocate network for instance. [ 1278.354042] env[61995]: DEBUG nova.network.neutron [-] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1278.473090] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.473444] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.473685] env[61995]: DEBUG nova.objects.instance [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'resources' on Instance uuid 47db0510-411f-415c-99a4-63ed1bcbcbc9 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1278.857053] env[61995]: INFO nova.compute.manager [-] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Took 1.21 seconds to deallocate network for instance. [ 1279.011830] env[61995]: DEBUG nova.compute.manager [req-7bbcb75e-31fa-431d-a380-cc5966691fdd req-1e097655-14b5-4eac-a9e0-42ef2ecc52e1 service nova] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Received event network-vif-deleted-be5b1c30-9807-4ce6-b41b-6e571c083675 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1279.029889] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0ac07f-228f-436d-b50b-b4e6e472398c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.037379] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1374d3-9b07-4b4c-8c89-ff146aa60a54 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.066285] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1889d067-8bf3-4e12-8afd-0ba3118c92fb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.073165] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37e1151-8e8f-4d7d-b5fd-d91687730fcc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.085810] env[61995]: DEBUG nova.compute.provider_tree [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1279.363457] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.589034] env[61995]: DEBUG nova.scheduler.client.report [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1280.093749] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.097123] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.733s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.097380] env[61995]: DEBUG nova.objects.instance [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lazy-loading 'resources' on Instance uuid 95af4a7a-bfe9-4bad-bb61-1df3736f81dc {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1280.118814] env[61995]: INFO nova.scheduler.client.report [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted allocations for instance 47db0510-411f-415c-99a4-63ed1bcbcbc9 [ 1280.627957] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6e187fad-34bc-4844-99ba-600d8ffbdbc4 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "47db0510-411f-415c-99a4-63ed1bcbcbc9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.783s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.645270] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172d6196-a82f-48f1-970a-11483c36c37c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.653994] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195ea26c-22e0-4a28-afa5-b1fd0800ba19 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.687168] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6c71cd-04cd-431d-99c2-369dfc664ef7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.695308] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72108a86-fead-4446-8477-4695b4d3cba2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.709613] env[61995]: DEBUG nova.compute.provider_tree [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1281.213199] env[61995]: DEBUG nova.scheduler.client.report [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1281.719137] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.737540] env[61995]: INFO nova.scheduler.client.report [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Deleted allocations for instance 95af4a7a-bfe9-4bad-bb61-1df3736f81dc [ 1281.830600] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.830901] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.245635] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c1bc2bfd-e01a-4147-952b-eb66c2d04d50 tempest-ServerTagsTestJSON-434499414 tempest-ServerTagsTestJSON-434499414-project-member] Lock "95af4a7a-bfe9-4bad-bb61-1df3736f81dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.787s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1282.333541] env[61995]: DEBUG nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1282.853985] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.854264] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.855701] env[61995]: INFO nova.compute.claims [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1283.906257] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfc3e7c-6618-4601-b48d-d70bdcbcc553 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.913771] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5181d3ef-2945-47ab-9761-ce23e674202b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.942747] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ee750a-343a-4659-ab75-7c9e61978cb3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.950075] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5769ca71-0433-4349-91ef-2c0c6fb309a4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.964434] env[61995]: DEBUG nova.compute.provider_tree [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1284.467844] env[61995]: DEBUG nova.scheduler.client.report [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1284.973222] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.973794] env[61995]: DEBUG nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1285.479486] env[61995]: DEBUG nova.compute.utils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1285.480953] env[61995]: DEBUG nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1285.481141] env[61995]: DEBUG nova.network.neutron [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1285.527445] env[61995]: DEBUG nova.policy [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e2f49fe94af402abe465e8415b16f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f3d88cb0eb54bc1a852c5f43f3806d9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1285.774688] env[61995]: DEBUG nova.network.neutron [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Successfully created port: 4ba4b083-dbc8-4645-900d-9ea4d14ce0da {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1285.984269] env[61995]: DEBUG nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1286.995637] env[61995]: DEBUG nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1287.020557] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1287.021167] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1287.021167] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1287.021167] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1287.021322] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1287.021469] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1287.021682] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1287.021847] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1287.022031] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1287.022221] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1287.022408] env[61995]: DEBUG nova.virt.hardware [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1287.023276] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9e9665-fa32-4fe5-b0ed-3fb1287a2cde {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.031069] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609d8c38-7c8e-4586-bd42-7888063c7a06 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.131591] env[61995]: DEBUG nova.compute.manager [req-e1ae2d64-766f-4044-b07f-bcb5109d395a req-05d431f6-59c9-40c4-9c12-588d31b54a6c service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Received event network-vif-plugged-4ba4b083-dbc8-4645-900d-9ea4d14ce0da {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1287.131901] env[61995]: DEBUG oslo_concurrency.lockutils [req-e1ae2d64-766f-4044-b07f-bcb5109d395a req-05d431f6-59c9-40c4-9c12-588d31b54a6c service nova] Acquiring lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1287.132227] env[61995]: DEBUG oslo_concurrency.lockutils [req-e1ae2d64-766f-4044-b07f-bcb5109d395a req-05d431f6-59c9-40c4-9c12-588d31b54a6c service nova] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1287.132450] env[61995]: DEBUG oslo_concurrency.lockutils [req-e1ae2d64-766f-4044-b07f-bcb5109d395a req-05d431f6-59c9-40c4-9c12-588d31b54a6c service nova] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1287.132732] env[61995]: DEBUG nova.compute.manager [req-e1ae2d64-766f-4044-b07f-bcb5109d395a req-05d431f6-59c9-40c4-9c12-588d31b54a6c service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] No waiting events found dispatching network-vif-plugged-4ba4b083-dbc8-4645-900d-9ea4d14ce0da {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1287.132958] env[61995]: WARNING nova.compute.manager [req-e1ae2d64-766f-4044-b07f-bcb5109d395a req-05d431f6-59c9-40c4-9c12-588d31b54a6c service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Received unexpected event network-vif-plugged-4ba4b083-dbc8-4645-900d-9ea4d14ce0da for instance with vm_state building and task_state spawning. [ 1287.212483] env[61995]: DEBUG nova.network.neutron [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Successfully updated port: 4ba4b083-dbc8-4645-900d-9ea4d14ce0da {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1287.715097] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-670c50f0-fd9b-4eb5-85c0-6dad9056b16c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1287.715178] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-670c50f0-fd9b-4eb5-85c0-6dad9056b16c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.715287] env[61995]: DEBUG nova.network.neutron [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1288.250048] env[61995]: DEBUG nova.network.neutron [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1288.369168] env[61995]: DEBUG nova.network.neutron [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Updating instance_info_cache with network_info: [{"id": "4ba4b083-dbc8-4645-900d-9ea4d14ce0da", "address": "fa:16:3e:81:cb:a1", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ba4b083-db", "ovs_interfaceid": "4ba4b083-dbc8-4645-900d-9ea4d14ce0da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.871722] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-670c50f0-fd9b-4eb5-85c0-6dad9056b16c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1288.872089] env[61995]: DEBUG nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Instance network_info: |[{"id": "4ba4b083-dbc8-4645-900d-9ea4d14ce0da", "address": "fa:16:3e:81:cb:a1", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ba4b083-db", "ovs_interfaceid": "4ba4b083-dbc8-4645-900d-9ea4d14ce0da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1288.872526] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:cb:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aef08290-001a-4ae8-aff0-1889e2211389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ba4b083-dbc8-4645-900d-9ea4d14ce0da', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1288.879880] env[61995]: DEBUG oslo.service.loopingcall [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1288.880103] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1288.880333] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5dd7e950-e9e4-4dad-8cd9-424b3edeeeb0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.901702] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1288.901702] env[61995]: value = "task-795579" [ 1288.901702] env[61995]: _type = "Task" [ 1288.901702] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.909202] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795579, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.159027] env[61995]: DEBUG nova.compute.manager [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Received event network-changed-4ba4b083-dbc8-4645-900d-9ea4d14ce0da {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1289.159280] env[61995]: DEBUG nova.compute.manager [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Refreshing instance network info cache due to event network-changed-4ba4b083-dbc8-4645-900d-9ea4d14ce0da. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1289.159465] env[61995]: DEBUG oslo_concurrency.lockutils [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] Acquiring lock "refresh_cache-670c50f0-fd9b-4eb5-85c0-6dad9056b16c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1289.159620] env[61995]: DEBUG oslo_concurrency.lockutils [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] Acquired lock "refresh_cache-670c50f0-fd9b-4eb5-85c0-6dad9056b16c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.159785] env[61995]: DEBUG nova.network.neutron [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Refreshing network info cache for port 4ba4b083-dbc8-4645-900d-9ea4d14ce0da {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1289.411494] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795579, 'name': CreateVM_Task, 'duration_secs': 0.293644} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.411764] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1289.418329] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1289.418517] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.418849] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1289.419098] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ac6dcd2-1403-46c2-a102-0271776f1de1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.423446] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1289.423446] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d9bc25-4b00-8d70-4d8e-e31fd1b007f4" [ 1289.423446] env[61995]: _type = "Task" [ 1289.423446] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.430541] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d9bc25-4b00-8d70-4d8e-e31fd1b007f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.843472] env[61995]: DEBUG nova.network.neutron [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Updated VIF entry in instance network info cache for port 4ba4b083-dbc8-4645-900d-9ea4d14ce0da. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1289.843847] env[61995]: DEBUG nova.network.neutron [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Updating instance_info_cache with network_info: [{"id": "4ba4b083-dbc8-4645-900d-9ea4d14ce0da", "address": "fa:16:3e:81:cb:a1", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ba4b083-db", "ovs_interfaceid": "4ba4b083-dbc8-4645-900d-9ea4d14ce0da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1289.934092] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d9bc25-4b00-8d70-4d8e-e31fd1b007f4, 'name': SearchDatastore_Task, 'duration_secs': 0.011348} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.934225] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.934417] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1289.934710] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1289.934908] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.935157] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1289.935456] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab8daffe-3b2b-499c-946c-dabb81c2b1e2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.943400] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1289.943622] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1289.944340] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a877e34-6251-45c9-bcfd-0ba97e9db81d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.949285] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1289.949285] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52363086-b58e-8fb0-fdcf-26da40b5bece" [ 1289.949285] env[61995]: _type = "Task" [ 1289.949285] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.956889] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52363086-b58e-8fb0-fdcf-26da40b5bece, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.347686] env[61995]: DEBUG oslo_concurrency.lockutils [req-f0688746-5f14-4d63-9ec2-96f8a96a770f req-bcb6a341-b955-4523-9b22-5c048563d2c5 service nova] Releasing lock "refresh_cache-670c50f0-fd9b-4eb5-85c0-6dad9056b16c" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1290.459270] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52363086-b58e-8fb0-fdcf-26da40b5bece, 'name': SearchDatastore_Task, 'duration_secs': 0.009576} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.460018] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc40d9a6-745c-459f-833b-d03608cef0de {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.464728] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1290.464728] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52d4773f-93bc-e3b3-d1da-c53302e02e14" [ 1290.464728] env[61995]: _type = "Task" [ 1290.464728] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.471787] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d4773f-93bc-e3b3-d1da-c53302e02e14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.977128] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52d4773f-93bc-e3b3-d1da-c53302e02e14, 'name': SearchDatastore_Task, 'duration_secs': 0.009161} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.977550] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1290.977980] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 670c50f0-fd9b-4eb5-85c0-6dad9056b16c/670c50f0-fd9b-4eb5-85c0-6dad9056b16c.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1290.978372] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad4a0ec5-e1eb-462a-b320-5a6c055248c3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.986170] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1290.986170] env[61995]: value = "task-795580" [ 1290.986170] env[61995]: _type = "Task" [ 1290.986170] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.993648] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.495810] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434685} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.496174] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore1] 670c50f0-fd9b-4eb5-85c0-6dad9056b16c/670c50f0-fd9b-4eb5-85c0-6dad9056b16c.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1291.496324] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1291.496571] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-392c3cd0-7675-43f0-a320-c22c26705af5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.502736] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1291.502736] env[61995]: value = "task-795581" [ 1291.502736] env[61995]: _type = "Task" [ 1291.502736] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.509739] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.013501] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061546} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.013759] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1292.014513] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8dd618-4d8e-41bd-8cac-cd832146611b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.035259] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] 670c50f0-fd9b-4eb5-85c0-6dad9056b16c/670c50f0-fd9b-4eb5-85c0-6dad9056b16c.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1292.035479] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b23b09b0-3348-413e-a958-77400f0f181c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.054073] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1292.054073] env[61995]: value = "task-795582" [ 1292.054073] env[61995]: _type = "Task" [ 1292.054073] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.061171] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795582, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.563974] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795582, 'name': ReconfigVM_Task, 'duration_secs': 0.309736} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.564398] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Reconfigured VM instance instance-00000075 to attach disk [datastore1] 670c50f0-fd9b-4eb5-85c0-6dad9056b16c/670c50f0-fd9b-4eb5-85c0-6dad9056b16c.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1292.564995] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2af35e4-bcb7-4156-ab3b-32da818e1598 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.571501] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1292.571501] env[61995]: value = "task-795583" [ 1292.571501] env[61995]: _type = "Task" [ 1292.571501] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.578654] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795583, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.081380] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795583, 'name': Rename_Task, 'duration_secs': 0.15139} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.081656] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1293.081894] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1eb3a2c-fd6d-401a-b1c2-fdc3e26ce0ca {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.088246] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1293.088246] env[61995]: value = "task-795584" [ 1293.088246] env[61995]: _type = "Task" [ 1293.088246] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.095362] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795584, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.597941] env[61995]: DEBUG oslo_vmware.api [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795584, 'name': PowerOnVM_Task, 'duration_secs': 0.468571} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.598397] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1293.599198] env[61995]: INFO nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Took 6.60 seconds to spawn the instance on the hypervisor. [ 1293.599198] env[61995]: DEBUG nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1293.599367] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1000c7-fd77-4362-8e3d-d3cc5922a178 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.118668] env[61995]: INFO nova.compute.manager [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Took 11.28 seconds to build instance. [ 1294.621332] env[61995]: DEBUG oslo_concurrency.lockutils [None req-1499df88-08f1-4217-969d-7aa9d767c745 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.790s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.086885] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce3163f-f551-4c67-ad92-ef2ec01ff332 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.093492] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a04ae553-52f0-4b7b-ae89-955b3d34207f tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Suspending the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1295.093731] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-937834fa-7adb-435b-a324-b8072c20dd40 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.099791] env[61995]: DEBUG oslo_vmware.api [None req-a04ae553-52f0-4b7b-ae89-955b3d34207f tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1295.099791] env[61995]: value = "task-795585" [ 1295.099791] env[61995]: _type = "Task" [ 1295.099791] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.107474] env[61995]: DEBUG oslo_vmware.api [None req-a04ae553-52f0-4b7b-ae89-955b3d34207f tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795585, 'name': SuspendVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.612197] env[61995]: DEBUG oslo_vmware.api [None req-a04ae553-52f0-4b7b-ae89-955b3d34207f tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795585, 'name': SuspendVM_Task} progress is 70%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.110150] env[61995]: DEBUG oslo_vmware.api [None req-a04ae553-52f0-4b7b-ae89-955b3d34207f tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795585, 'name': SuspendVM_Task, 'duration_secs': 0.595254} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.110556] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-a04ae553-52f0-4b7b-ae89-955b3d34207f tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Suspended the VM {{(pid=61995) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1296.110609] env[61995]: DEBUG nova.compute.manager [None req-a04ae553-52f0-4b7b-ae89-955b3d34207f tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1296.111364] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66222333-69d6-433a-b63a-d646b49698da {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.465985] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1297.466361] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.466569] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1297.466763] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.466938] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1297.469136] env[61995]: INFO nova.compute.manager [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Terminating instance [ 1297.470889] env[61995]: DEBUG nova.compute.manager [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1297.471102] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1297.471914] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16217f54-943e-442d-8c37-2498b1349d92 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.479713] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1297.479933] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0c4cedc-c7ae-49e9-b0e8-3b2326c9a629 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.549750] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1297.549974] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Deleting contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1297.550186] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleting the datastore file [datastore1] 670c50f0-fd9b-4eb5-85c0-6dad9056b16c {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1297.550444] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e14ce552-ba79-4f98-aa05-f0259ecb5b14 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.556962] env[61995]: DEBUG oslo_vmware.api [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1297.556962] env[61995]: value = "task-795587" [ 1297.556962] env[61995]: _type = "Task" [ 1297.556962] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.564461] env[61995]: DEBUG oslo_vmware.api [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.066238] env[61995]: DEBUG oslo_vmware.api [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146161} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.066440] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1298.066633] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Deleted contents of the VM from datastore datastore1 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1298.066811] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1298.066985] env[61995]: INFO nova.compute.manager [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1298.067255] env[61995]: DEBUG oslo.service.loopingcall [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1298.067454] env[61995]: DEBUG nova.compute.manager [-] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1298.067548] env[61995]: DEBUG nova.network.neutron [-] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1298.340567] env[61995]: DEBUG nova.compute.manager [req-a4fee2ce-441c-4762-bb9b-05acbd1debb5 req-26f110b1-97f8-4cbf-b022-20c6781feac1 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Received event network-vif-deleted-4ba4b083-dbc8-4645-900d-9ea4d14ce0da {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1298.340781] env[61995]: INFO nova.compute.manager [req-a4fee2ce-441c-4762-bb9b-05acbd1debb5 req-26f110b1-97f8-4cbf-b022-20c6781feac1 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Neutron deleted interface 4ba4b083-dbc8-4645-900d-9ea4d14ce0da; detaching it from the instance and deleting it from the info cache [ 1298.340954] env[61995]: DEBUG nova.network.neutron [req-a4fee2ce-441c-4762-bb9b-05acbd1debb5 req-26f110b1-97f8-4cbf-b022-20c6781feac1 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.819211] env[61995]: DEBUG nova.network.neutron [-] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.843289] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91cfce1c-42a0-4827-b13d-607c3cc24b7f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.853168] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d282678-de7b-4340-9911-138b7a4bd1d4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.878578] env[61995]: DEBUG nova.compute.manager [req-a4fee2ce-441c-4762-bb9b-05acbd1debb5 req-26f110b1-97f8-4cbf-b022-20c6781feac1 service nova] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Detach interface failed, port_id=4ba4b083-dbc8-4645-900d-9ea4d14ce0da, reason: Instance 670c50f0-fd9b-4eb5-85c0-6dad9056b16c could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1299.322359] env[61995]: INFO nova.compute.manager [-] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Took 1.25 seconds to deallocate network for instance. [ 1299.829465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1299.829875] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.829964] env[61995]: DEBUG nova.objects.instance [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'resources' on Instance uuid 670c50f0-fd9b-4eb5-85c0-6dad9056b16c {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1300.373098] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb04fdc9-7b3a-4bf1-95bc-328b8319a8f9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.380558] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e214e78-5bc4-4742-ac62-0cc17859854a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.409966] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4527f98f-01ee-440b-a193-2f7737cf341e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.416810] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c729005c-e5e3-4819-8516-f5d04d3abd82 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.429353] env[61995]: DEBUG nova.compute.provider_tree [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1300.933291] env[61995]: DEBUG nova.scheduler.client.report [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1301.437159] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.455764] env[61995]: INFO nova.scheduler.client.report [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted allocations for instance 670c50f0-fd9b-4eb5-85c0-6dad9056b16c [ 1301.964646] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f45ade62-6ad8-4d3c-84ee-d3405e48a1f6 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "670c50f0-fd9b-4eb5-85c0-6dad9056b16c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.498s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1303.334543] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1303.334840] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1303.837532] env[61995]: DEBUG nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1304.360118] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1304.360451] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1304.361894] env[61995]: INFO nova.compute.claims [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1305.404568] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dd1011-f361-450f-8caa-904faa7b9a26 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.411787] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f4283f-ca0f-4747-829e-626b9409effe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.442444] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9264052a-0054-4e29-b1cd-493dfe35bd48 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.449780] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7931f48e-9337-4c14-b7f3-884a8ea5d8f3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.462534] env[61995]: DEBUG nova.compute.provider_tree [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1305.965793] env[61995]: DEBUG nova.scheduler.client.report [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1306.471465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.472034] env[61995]: DEBUG nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1306.976795] env[61995]: DEBUG nova.compute.utils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1306.978358] env[61995]: DEBUG nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1306.978529] env[61995]: DEBUG nova.network.neutron [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1307.029462] env[61995]: DEBUG nova.policy [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e2f49fe94af402abe465e8415b16f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f3d88cb0eb54bc1a852c5f43f3806d9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1307.274226] env[61995]: DEBUG nova.network.neutron [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Successfully created port: cacfdcc7-4c0e-4623-80f0-0221590ba560 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1307.482093] env[61995]: DEBUG nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1308.491489] env[61995]: DEBUG nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1308.516588] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1308.516881] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1308.516993] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1308.517193] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1308.517344] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1308.517493] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1308.517752] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1308.517990] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1308.518215] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1308.518399] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1308.518578] env[61995]: DEBUG nova.virt.hardware [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1308.519495] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80ad0c7-c4d7-4ccc-8fca-7e68907272d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.527318] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee654ae5-c359-4bb4-8ab4-215ed74962e9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.628343] env[61995]: DEBUG nova.compute.manager [req-7a9f7717-668e-4189-b69d-deaf435b3153 req-b76f031b-ad89-4913-8afb-fab0f631e519 service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Received event network-vif-plugged-cacfdcc7-4c0e-4623-80f0-0221590ba560 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1308.628616] env[61995]: DEBUG oslo_concurrency.lockutils [req-7a9f7717-668e-4189-b69d-deaf435b3153 req-b76f031b-ad89-4913-8afb-fab0f631e519 service nova] Acquiring lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1308.628783] env[61995]: DEBUG oslo_concurrency.lockutils [req-7a9f7717-668e-4189-b69d-deaf435b3153 req-b76f031b-ad89-4913-8afb-fab0f631e519 service nova] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1308.628956] env[61995]: DEBUG oslo_concurrency.lockutils [req-7a9f7717-668e-4189-b69d-deaf435b3153 req-b76f031b-ad89-4913-8afb-fab0f631e519 service nova] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1308.629206] env[61995]: DEBUG nova.compute.manager [req-7a9f7717-668e-4189-b69d-deaf435b3153 req-b76f031b-ad89-4913-8afb-fab0f631e519 service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] No waiting events found dispatching network-vif-plugged-cacfdcc7-4c0e-4623-80f0-0221590ba560 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1308.629438] env[61995]: WARNING nova.compute.manager [req-7a9f7717-668e-4189-b69d-deaf435b3153 req-b76f031b-ad89-4913-8afb-fab0f631e519 service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Received unexpected event network-vif-plugged-cacfdcc7-4c0e-4623-80f0-0221590ba560 for instance with vm_state building and task_state spawning. [ 1308.709077] env[61995]: DEBUG nova.network.neutron [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Successfully updated port: cacfdcc7-4c0e-4623-80f0-0221590ba560 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1309.211460] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1309.211672] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.211834] env[61995]: DEBUG nova.network.neutron [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1309.734994] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1309.735439] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1309.744412] env[61995]: DEBUG nova.network.neutron [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1309.875939] env[61995]: DEBUG nova.network.neutron [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [{"id": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "address": "fa:16:3e:c5:69:4f", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcacfdcc7-4c", "ovs_interfaceid": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.238376] env[61995]: INFO nova.compute.manager [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Detaching volume 9834a3bf-0fe3-4814-8d27-5aadd4a878ed [ 1310.269559] env[61995]: INFO nova.virt.block_device [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Attempting to driver detach volume 9834a3bf-0fe3-4814-8d27-5aadd4a878ed from mountpoint /dev/sdb [ 1310.269811] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Volume detach. Driver type: vmdk {{(pid=61995) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1310.270008] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185457', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'name': 'volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '24853688-b972-48e6-be7b-7f0cf634c60d', 'attached_at': '', 'detached_at': '', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'serial': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1310.270914] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a8e478-e176-4465-ace7-3bda4bc11b8a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.292125] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ed6b03-72c2-4801-8a5f-b61ab1d53a7b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.299019] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34323522-bd87-4dff-bf6a-bcdbe2793f15 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.318636] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e938f4-906c-44cb-8375-245243417e6f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.334153] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] The volume has not been displaced from its original location: [datastore2] volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed/volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed.vmdk. No consolidation needed. {{(pid=61995) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1310.339392] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1310.339664] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63e060a1-b041-4576-9df2-b3fe1449265d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.357842] env[61995]: DEBUG oslo_vmware.api [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1310.357842] env[61995]: value = "task-795588" [ 1310.357842] env[61995]: _type = "Task" [ 1310.357842] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.365430] env[61995]: DEBUG oslo_vmware.api [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.377983] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1310.378292] env[61995]: DEBUG nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Instance network_info: |[{"id": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "address": "fa:16:3e:c5:69:4f", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcacfdcc7-4c", "ovs_interfaceid": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1310.378668] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:69:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aef08290-001a-4ae8-aff0-1889e2211389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cacfdcc7-4c0e-4623-80f0-0221590ba560', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1310.385822] env[61995]: DEBUG oslo.service.loopingcall [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1310.386039] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1310.386274] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a5fe1a8-8e58-4fb8-beb7-5f91c13e35f8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.404812] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1310.404812] env[61995]: value = "task-795589" [ 1310.404812] env[61995]: _type = "Task" [ 1310.404812] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.412131] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795589, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.653874] env[61995]: DEBUG nova.compute.manager [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Received event network-changed-cacfdcc7-4c0e-4623-80f0-0221590ba560 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1310.654099] env[61995]: DEBUG nova.compute.manager [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Refreshing instance network info cache due to event network-changed-cacfdcc7-4c0e-4623-80f0-0221590ba560. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1310.654321] env[61995]: DEBUG oslo_concurrency.lockutils [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] Acquiring lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1310.654473] env[61995]: DEBUG oslo_concurrency.lockutils [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] Acquired lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1310.654636] env[61995]: DEBUG nova.network.neutron [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Refreshing network info cache for port cacfdcc7-4c0e-4623-80f0-0221590ba560 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1310.866884] env[61995]: DEBUG oslo_vmware.api [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795588, 'name': ReconfigVM_Task, 'duration_secs': 0.218812} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.867191] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1310.871716] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbdc6d10-6e60-4fde-9a49-8323808896f6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.886546] env[61995]: DEBUG oslo_vmware.api [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1310.886546] env[61995]: value = "task-795590" [ 1310.886546] env[61995]: _type = "Task" [ 1310.886546] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.894017] env[61995]: DEBUG oslo_vmware.api [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.915718] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795589, 'name': CreateVM_Task, 'duration_secs': 0.305227} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.915879] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1310.916544] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1310.916711] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1310.917037] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1310.917275] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f414692b-0e11-425f-9ec0-91f27e2e8da6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.921562] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1310.921562] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52b5af8f-eb22-900d-13c9-9998b8569ead" [ 1310.921562] env[61995]: _type = "Task" [ 1310.921562] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.929043] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b5af8f-eb22-900d-13c9-9998b8569ead, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.333850] env[61995]: DEBUG nova.network.neutron [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updated VIF entry in instance network info cache for port cacfdcc7-4c0e-4623-80f0-0221590ba560. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1311.334242] env[61995]: DEBUG nova.network.neutron [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [{"id": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "address": "fa:16:3e:c5:69:4f", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcacfdcc7-4c", "ovs_interfaceid": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1311.395611] env[61995]: DEBUG oslo_vmware.api [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795590, 'name': ReconfigVM_Task, 'duration_secs': 0.129502} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.395930] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-185457', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'name': 'volume-9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '24853688-b972-48e6-be7b-7f0cf634c60d', 'attached_at': '', 'detached_at': '', 'volume_id': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed', 'serial': '9834a3bf-0fe3-4814-8d27-5aadd4a878ed'} {{(pid=61995) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1311.431674] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52b5af8f-eb22-900d-13c9-9998b8569ead, 'name': SearchDatastore_Task, 'duration_secs': 0.011922} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.431978] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.432236] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1311.432474] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1311.432625] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1311.432806] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1311.433072] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37e6b34a-4f3f-45e4-9004-efa7965a27e7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.441947] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1311.442151] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1311.442840] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0969762b-a279-41ca-8155-0599e7f13eac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.447850] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1311.447850] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ec8e9b-ceea-a82e-5359-c9c80da5ee83" [ 1311.447850] env[61995]: _type = "Task" [ 1311.447850] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.455314] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ec8e9b-ceea-a82e-5359-c9c80da5ee83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.837333] env[61995]: DEBUG oslo_concurrency.lockutils [req-1af7a307-7d51-4d50-b3af-b4c8fcfccdc9 req-8e9ced80-396f-4155-b0fb-21295e29d5ec service nova] Releasing lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.935488] env[61995]: DEBUG nova.objects.instance [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'flavor' on Instance uuid 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1311.958682] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ec8e9b-ceea-a82e-5359-c9c80da5ee83, 'name': SearchDatastore_Task, 'duration_secs': 0.008544} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.959486] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d39d4a2c-4fb5-416d-8b5c-43236a936a8e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.965134] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1311.965134] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c24705-e5b4-086b-662e-6628fb1e7ffe" [ 1311.965134] env[61995]: _type = "Task" [ 1311.965134] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.974252] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c24705-e5b4-086b-662e-6628fb1e7ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.474716] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c24705-e5b4-086b-662e-6628fb1e7ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.011192} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.474975] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1312.475253] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] ab8be459-0fb7-4f8d-ba53-66a7d64b30fc/ab8be459-0fb7-4f8d-ba53-66a7d64b30fc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1312.475494] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d6761cc-b257-4639-ae45-34b4cfc1eea6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.482393] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1312.482393] env[61995]: value = "task-795591" [ 1312.482393] env[61995]: _type = "Task" [ 1312.482393] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.489409] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795591, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.942720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-f69b20f7-13f6-43a0-a4cb-98c4d5dca0d1 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.207s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.991873] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795591, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.398452} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.992144] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] ab8be459-0fb7-4f8d-ba53-66a7d64b30fc/ab8be459-0fb7-4f8d-ba53-66a7d64b30fc.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1312.992359] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1312.992602] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e60d9866-17a4-4b91-9363-57d2e4d68af2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.000619] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1313.000619] env[61995]: value = "task-795592" [ 1313.000619] env[61995]: _type = "Task" [ 1313.000619] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.008073] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795592, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.510128] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795592, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059866} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.510658] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1313.511411] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f308e2-e690-4132-bf37-45dbde6812d8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.532898] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] ab8be459-0fb7-4f8d-ba53-66a7d64b30fc/ab8be459-0fb7-4f8d-ba53-66a7d64b30fc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1313.533170] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d48c4977-b0ce-4ada-92c5-3c695084fc48 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.552167] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1313.552167] env[61995]: value = "task-795593" [ 1313.552167] env[61995]: _type = "Task" [ 1313.552167] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.559763] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795593, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.979555] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.979932] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.980043] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.980239] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.980416] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.982838] env[61995]: INFO nova.compute.manager [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Terminating instance [ 1313.984583] env[61995]: DEBUG nova.compute.manager [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1313.984775] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1313.985614] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e44a52e-ad2e-43ca-86f6-a338e7be0af6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.992821] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1313.993044] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2f2f086-884b-44cd-9ab6-30c5a87a6355 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.998842] env[61995]: DEBUG oslo_vmware.api [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1313.998842] env[61995]: value = "task-795594" [ 1313.998842] env[61995]: _type = "Task" [ 1313.998842] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.006179] env[61995]: DEBUG oslo_vmware.api [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.064737] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795593, 'name': ReconfigVM_Task, 'duration_secs': 0.25864} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.065197] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Reconfigured VM instance instance-00000076 to attach disk [datastore2] ab8be459-0fb7-4f8d-ba53-66a7d64b30fc/ab8be459-0fb7-4f8d-ba53-66a7d64b30fc.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1314.066014] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-495e19ea-8933-4967-9379-e2af143255d7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.073462] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1314.073462] env[61995]: value = "task-795595" [ 1314.073462] env[61995]: _type = "Task" [ 1314.073462] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.083303] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795595, 'name': Rename_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.508494] env[61995]: DEBUG oslo_vmware.api [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795594, 'name': PowerOffVM_Task, 'duration_secs': 0.182224} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.508749] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1314.508919] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1314.509526] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac48918e-4cbe-47cf-9fad-ac8242642683 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.583257] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795595, 'name': Rename_Task, 'duration_secs': 0.132417} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.584302] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1314.584581] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1314.584753] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1314.584927] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleting the datastore file [datastore2] 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1314.585162] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fce37b8b-ab35-44a8-afa3-4a47fec107d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.586563] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09bafec4-8985-4364-b11b-5de1554a080c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.592783] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1314.592783] env[61995]: value = "task-795597" [ 1314.592783] env[61995]: _type = "Task" [ 1314.592783] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.596578] env[61995]: DEBUG oslo_vmware.api [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1314.596578] env[61995]: value = "task-795598" [ 1314.596578] env[61995]: _type = "Task" [ 1314.596578] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.603088] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.605850] env[61995]: DEBUG oslo_vmware.api [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795598, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.103395] env[61995]: DEBUG oslo_vmware.api [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795597, 'name': PowerOnVM_Task, 'duration_secs': 0.427631} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.104032] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1315.105521] env[61995]: INFO nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Took 6.61 seconds to spawn the instance on the hypervisor. [ 1315.105521] env[61995]: DEBUG nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1315.105521] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f59348-34a1-4f34-a8cb-d8328e0c1a6a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.110131] env[61995]: DEBUG oslo_vmware.api [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795598, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13121} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.110678] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1315.110894] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1315.111114] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1315.111304] env[61995]: INFO nova.compute.manager [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1315.111543] env[61995]: DEBUG oslo.service.loopingcall [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1315.111731] env[61995]: DEBUG nova.compute.manager [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1315.111851] env[61995]: DEBUG nova.network.neutron [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1315.606859] env[61995]: DEBUG nova.compute.manager [req-306c600d-eebb-4feb-8fb0-1f494cabcac2 req-ededec0f-60f6-4b9b-83c9-a599a1ab5d80 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Received event network-vif-deleted-47ddce51-cfec-49c4-ba3b-b34c86140652 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1315.607141] env[61995]: INFO nova.compute.manager [req-306c600d-eebb-4feb-8fb0-1f494cabcac2 req-ededec0f-60f6-4b9b-83c9-a599a1ab5d80 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Neutron deleted interface 47ddce51-cfec-49c4-ba3b-b34c86140652; detaching it from the instance and deleting it from the info cache [ 1315.607390] env[61995]: DEBUG nova.network.neutron [req-306c600d-eebb-4feb-8fb0-1f494cabcac2 req-ededec0f-60f6-4b9b-83c9-a599a1ab5d80 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1315.625783] env[61995]: INFO nova.compute.manager [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Took 11.28 seconds to build instance. [ 1316.016502] env[61995]: DEBUG nova.network.neutron [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.110068] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53a84dd9-3203-4033-b977-60adaee1ed9d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.120397] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989affd3-0424-470b-b477-e3fd57847332 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.134518] env[61995]: DEBUG oslo_concurrency.lockutils [None req-31e71372-492c-4081-a0ae-6d90867e391b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.800s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1316.147956] env[61995]: DEBUG nova.compute.manager [req-306c600d-eebb-4feb-8fb0-1f494cabcac2 req-ededec0f-60f6-4b9b-83c9-a599a1ab5d80 service nova] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Detach interface failed, port_id=47ddce51-cfec-49c4-ba3b-b34c86140652, reason: Instance 24853688-b972-48e6-be7b-7f0cf634c60d could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1316.519299] env[61995]: INFO nova.compute.manager [-] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Took 1.41 seconds to deallocate network for instance. [ 1317.025380] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.025666] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1317.025898] env[61995]: DEBUG nova.objects.instance [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'resources' on Instance uuid 24853688-b972-48e6-be7b-7f0cf634c60d {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1317.250064] env[61995]: DEBUG nova.compute.manager [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Stashing vm_state: active {{(pid=61995) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1317.576194] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9095c59-6300-4314-a956-e58ea9ab7241 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.584102] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4b52dc-98b5-4078-a9df-b16e2332910d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.612726] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372d6085-1022-42c0-a878-65fdffc987cf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.619925] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9653e73a-b04e-421e-a0d9-ec6df980981d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.634202] env[61995]: DEBUG nova.compute.provider_tree [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1317.766909] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1318.136982] env[61995]: DEBUG nova.scheduler.client.report [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1318.642424] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1318.644483] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.878s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1318.663633] env[61995]: INFO nova.scheduler.client.report [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted allocations for instance 24853688-b972-48e6-be7b-7f0cf634c60d [ 1319.149630] env[61995]: INFO nova.compute.claims [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1319.170458] env[61995]: DEBUG oslo_concurrency.lockutils [None req-125aba8d-fa39-4a15-9106-bea1a60e0aa7 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "24853688-b972-48e6-be7b-7f0cf634c60d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.190s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.656942] env[61995]: INFO nova.compute.resource_tracker [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating resource usage from migration edcaea7f-f25c-4f42-b053-8d4480db3721 [ 1319.701702] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b118014b-0a6c-41e8-92a3-b81a19fc2480 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.709454] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a187185a-fc5b-4ba5-be5d-de1c149a31bc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.739097] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd797a93-05ef-40c8-86cf-b06a733d009e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.746671] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61280a4-3689-4674-a5c4-7aa92b50b8b3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.760767] env[61995]: DEBUG nova.compute.provider_tree [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1320.265533] env[61995]: DEBUG nova.scheduler.client.report [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1320.770827] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.126s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1320.771171] env[61995]: INFO nova.compute.manager [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Migrating [ 1320.856755] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1320.857020] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1321.285120] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1321.285387] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.285542] env[61995]: DEBUG nova.network.neutron [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1321.359375] env[61995]: DEBUG nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Starting instance... {{(pid=61995) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1321.880592] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1321.880864] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1321.882387] env[61995]: INFO nova.compute.claims [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1321.993227] env[61995]: DEBUG nova.network.neutron [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [{"id": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "address": "fa:16:3e:c5:69:4f", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcacfdcc7-4c", "ovs_interfaceid": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1322.495719] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1322.937519] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b7d0d9-7e04-4c56-8803-5754b69ad5b5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.944778] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0860b9-a506-4e08-8eb3-c6a6308f4ee3 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.973184] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aee6daf-7537-4ff1-9443-653c82af7da8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.979880] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adce23dc-dfe4-4258-9587-0f5580b732b4 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.992290] env[61995]: DEBUG nova.compute.provider_tree [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1323.495717] env[61995]: DEBUG nova.scheduler.client.report [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1324.000347] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.000950] env[61995]: DEBUG nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Start building networks asynchronously for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1324.008413] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf274d3-60d6-4f97-938c-5327733858dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.027955] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance 'ab8be459-0fb7-4f8d-ba53-66a7d64b30fc' progress to 0 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1324.506241] env[61995]: DEBUG nova.compute.utils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Using /dev/sd instead of None {{(pid=61995) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1324.507724] env[61995]: DEBUG nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Allocating IP information in the background. {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1324.507911] env[61995]: DEBUG nova.network.neutron [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] allocate_for_instance() {{(pid=61995) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1324.534106] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1324.534400] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a6a99a1-31c4-4694-a03c-540b30bc0c4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.542579] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1324.542579] env[61995]: value = "task-795599" [ 1324.542579] env[61995]: _type = "Task" [ 1324.542579] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.547068] env[61995]: DEBUG nova.policy [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8feed88faa214a6fa5c51f6f0cca4ea7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49e78af0ad9340258211bf92e447021c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61995) authorize /opt/stack/nova/nova/policy.py:203}} [ 1324.553696] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795599, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.840194] env[61995]: DEBUG nova.network.neutron [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Successfully created port: 6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1325.011212] env[61995]: DEBUG nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Start building block device mappings for instance. {{(pid=61995) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1325.052205] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795599, 'name': PowerOffVM_Task, 'duration_secs': 0.2887} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.052518] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1325.052718] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance 'ab8be459-0fb7-4f8d-ba53-66a7d64b30fc' progress to 17 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1325.558978] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1325.559268] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1325.559431] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1325.559617] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1325.559768] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1325.559919] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1325.560147] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1325.560315] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1325.560490] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1325.560656] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1325.560832] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1325.565926] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05d1c367-818a-442f-b84e-2dacf945e9ef {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.582075] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1325.582075] env[61995]: value = "task-795600" [ 1325.582075] env[61995]: _type = "Task" [ 1325.582075] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.590209] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.021199] env[61995]: DEBUG nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Start spawning the instance on the hypervisor. {{(pid=61995) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1326.046765] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-10T08:51:56Z,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ae745f566b90403c8e615c7069e2827b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-10T08:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1326.047052] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1326.047220] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1326.047408] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1326.047560] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1326.047713] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1326.047928] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1326.048114] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1326.048320] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1326.048503] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1326.048683] env[61995]: DEBUG nova.virt.hardware [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1326.049587] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b8afd6-54b5-4535-8e67-f2cb2cee41a2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.057565] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5527479-c8e9-4b4e-a5f5-d2ead4ac7724 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.090054] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795600, 'name': ReconfigVM_Task, 'duration_secs': 0.167412} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.090352] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance 'ab8be459-0fb7-4f8d-ba53-66a7d64b30fc' progress to 33 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1326.222289] env[61995]: DEBUG nova.compute.manager [req-86bdd53b-875c-42e2-aaec-99c92a917819 req-d0fcf0c7-2b7d-49e2-8630-78173a00d54d service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-vif-plugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1326.222289] env[61995]: DEBUG oslo_concurrency.lockutils [req-86bdd53b-875c-42e2-aaec-99c92a917819 req-d0fcf0c7-2b7d-49e2-8630-78173a00d54d service nova] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.222289] env[61995]: DEBUG oslo_concurrency.lockutils [req-86bdd53b-875c-42e2-aaec-99c92a917819 req-d0fcf0c7-2b7d-49e2-8630-78173a00d54d service nova] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.222289] env[61995]: DEBUG oslo_concurrency.lockutils [req-86bdd53b-875c-42e2-aaec-99c92a917819 req-d0fcf0c7-2b7d-49e2-8630-78173a00d54d service nova] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.222289] env[61995]: DEBUG nova.compute.manager [req-86bdd53b-875c-42e2-aaec-99c92a917819 req-d0fcf0c7-2b7d-49e2-8630-78173a00d54d service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] No waiting events found dispatching network-vif-plugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1326.223386] env[61995]: WARNING nova.compute.manager [req-86bdd53b-875c-42e2-aaec-99c92a917819 req-d0fcf0c7-2b7d-49e2-8630-78173a00d54d service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received unexpected event network-vif-plugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 for instance with vm_state building and task_state spawning. [ 1326.343468] env[61995]: DEBUG nova.network.neutron [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Successfully updated port: 6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1326.597016] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c1739132-aff3-4644-9522-fc02cf765bc8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1326.597305] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1326.597469] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1326.597662] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1326.597814] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1326.597969] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1326.598211] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1326.598396] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1326.598576] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1326.598736] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1326.598915] env[61995]: DEBUG nova.virt.hardware [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1326.604935] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Reconfiguring VM instance instance-00000076 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1326.605250] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bac07266-c828-431e-854a-11d464dbbbbe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.624420] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1326.624420] env[61995]: value = "task-795601" [ 1326.624420] env[61995]: _type = "Task" [ 1326.624420] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.631913] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795601, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.846525] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1326.846647] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.846799] env[61995]: DEBUG nova.network.neutron [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1327.133946] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795601, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.376758] env[61995]: DEBUG nova.network.neutron [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Instance cache missing network info. {{(pid=61995) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1327.499759] env[61995]: DEBUG nova.network.neutron [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1327.635113] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795601, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.003027] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1328.003027] env[61995]: DEBUG nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Instance network_info: |[{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61995) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1328.003403] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:7c:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '28d04eee-6dbb-491a-a999-b659c799679d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ab1828b-41a9-448a-853b-3ceb09e93d80', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1328.010745] env[61995]: DEBUG oslo.service.loopingcall [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1328.010954] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1328.011197] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ef413e6-0be1-4aa4-8909-3c25e362c420 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.030487] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1328.030487] env[61995]: value = "task-795602" [ 1328.030487] env[61995]: _type = "Task" [ 1328.030487] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.037555] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795602, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.134942] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795601, 'name': ReconfigVM_Task, 'duration_secs': 1.170017} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.135328] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Reconfigured VM instance instance-00000076 to detach disk 2000 {{(pid=61995) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1328.136053] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6149c3-77fe-4c27-a6fe-2cbc6685ff72 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.156912] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] ab8be459-0fb7-4f8d-ba53-66a7d64b30fc/ab8be459-0fb7-4f8d-ba53-66a7d64b30fc.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1328.157176] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daca20f6-4669-4255-8e01-da6e7257d9fe {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.176076] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1328.176076] env[61995]: value = "task-795603" [ 1328.176076] env[61995]: _type = "Task" [ 1328.176076] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.183294] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795603, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.203867] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1328.204125] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1328.204260] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1328.204390] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1328.250975] env[61995]: DEBUG nova.compute.manager [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1328.251186] env[61995]: DEBUG nova.compute.manager [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing instance network info cache due to event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1328.251404] env[61995]: DEBUG oslo_concurrency.lockutils [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1328.251555] env[61995]: DEBUG oslo_concurrency.lockutils [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.251723] env[61995]: DEBUG nova.network.neutron [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1328.540997] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795602, 'name': CreateVM_Task, 'duration_secs': 0.303727} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.541182] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1328.541889] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1328.542078] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.542404] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1328.542651] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c4e0a02-7a45-4950-80f4-484a5dbc43a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.546756] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1328.546756] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52f5137b-88f2-ebe0-40f9-be20d60e51b5" [ 1328.546756] env[61995]: _type = "Task" [ 1328.546756] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.553765] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f5137b-88f2-ebe0-40f9-be20d60e51b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.685450] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795603, 'name': ReconfigVM_Task, 'duration_secs': 0.270547} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.685733] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Reconfigured VM instance instance-00000076 to attach disk [datastore2] ab8be459-0fb7-4f8d-ba53-66a7d64b30fc/ab8be459-0fb7-4f8d-ba53-66a7d64b30fc.vmdk or device None with type thin {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1328.686010] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance 'ab8be459-0fb7-4f8d-ba53-66a7d64b30fc' progress to 50 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1328.707674] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Skipping network cache update for instance because it is Building. {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1328.707957] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1328.708097] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.708220] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1328.708374] env[61995]: DEBUG nova.objects.instance [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lazy-loading 'info_cache' on Instance uuid ab8be459-0fb7-4f8d-ba53-66a7d64b30fc {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1328.927736] env[61995]: DEBUG nova.network.neutron [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updated VIF entry in instance network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1328.928118] env[61995]: DEBUG nova.network.neutron [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.056731] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52f5137b-88f2-ebe0-40f9-be20d60e51b5, 'name': SearchDatastore_Task, 'duration_secs': 0.009815} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.057040] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1329.057282] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Processing image c1739132-aff3-4644-9522-fc02cf765bc8 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1329.057524] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1329.057675] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1329.057861] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1329.058123] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f8a960c-dbb6-4e2e-90c2-71fc2ed18579 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.065829] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1329.065973] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1329.066882] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfd5a9b6-7212-42de-87e4-75256bb78888 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.071359] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1329.071359] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52c7645a-edb1-a3f5-2c7c-f896b7551f4c" [ 1329.071359] env[61995]: _type = "Task" [ 1329.071359] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.078100] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c7645a-edb1-a3f5-2c7c-f896b7551f4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.192496] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575c2aab-9acb-4f06-abc3-b28ea5941310 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.210331] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f103c01-24d7-4e94-82e9-e2190de054a7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.228327] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance 'ab8be459-0fb7-4f8d-ba53-66a7d64b30fc' progress to 67 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1329.430492] env[61995]: DEBUG oslo_concurrency.lockutils [req-5f11c2cb-4a0e-434c-b191-e8c82893db14 req-0a60c344-fadd-4e10-a5aa-85d714ce8844 service nova] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1329.584030] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52c7645a-edb1-a3f5-2c7c-f896b7551f4c, 'name': SearchDatastore_Task, 'duration_secs': 0.007362} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.584811] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1360566-cfe0-4097-b272-0806c88cf27c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.590153] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1329.590153] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52ef4114-d1c8-5b63-ffdf-969beb18bc12" [ 1329.590153] env[61995]: _type = "Task" [ 1329.590153] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.597253] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ef4114-d1c8-5b63-ffdf-969beb18bc12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.766727] env[61995]: DEBUG nova.network.neutron [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Port cacfdcc7-4c0e-4623-80f0-0221590ba560 binding to destination host cpu-1 is already ACTIVE {{(pid=61995) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1330.100597] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52ef4114-d1c8-5b63-ffdf-969beb18bc12, 'name': SearchDatastore_Task, 'duration_secs': 0.009997} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.100883] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1330.101180] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1330.101442] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ef8bed8-b8e2-4bd4-a707-d56350d6cdb8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.107609] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1330.107609] env[61995]: value = "task-795604" [ 1330.107609] env[61995]: _type = "Task" [ 1330.107609] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.114588] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.427573] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [{"id": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "address": "fa:16:3e:c5:69:4f", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcacfdcc7-4c", "ovs_interfaceid": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1330.617511] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795604, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430405} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.617723] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1739132-aff3-4644-9522-fc02cf765bc8/c1739132-aff3-4644-9522-fc02cf765bc8.vmdk to [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1330.617925] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Extending root virtual disk to 1048576 {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1330.618903] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a542ae88-c650-49f6-90df-87cec4b8f9c7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.624574] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1330.624574] env[61995]: value = "task-795605" [ 1330.624574] env[61995]: _type = "Task" [ 1330.624574] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.631851] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.790039] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.790039] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.790039] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.930272] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1330.930493] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1330.930691] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.931577] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.931577] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.931577] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.931577] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.931577] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.931577] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1330.931862] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.134202] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795605, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057461} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.134483] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Extended root virtual disk {{(pid=61995) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1331.135246] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58ca552-0ff0-413e-8154-4a2398e473a1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.157463] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1331.157686] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-896a2f63-6dac-409d-ae1d-762e9a995855 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.176568] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1331.176568] env[61995]: value = "task-795606" [ 1331.176568] env[61995]: _type = "Task" [ 1331.176568] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.184110] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795606, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.434863] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.435309] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.435309] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1331.435519] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1331.436377] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12699558-307e-4931-9359-524fad2b3980 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.444527] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d7cd56-c3f9-4886-8d60-b5b97f5f7dac {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.457742] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3e8f73-0524-4e17-b288-a716b7416672 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.463900] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323b2852-82e0-4c0b-9028-15b068d97ef5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.491011] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181411MB free_disk=66GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1331.491166] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.491351] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.686393] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795606, 'name': ReconfigVM_Task, 'duration_secs': 0.256656} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.686640] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk or device None with type sparse {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1331.687282] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4796969a-c006-4e45-942c-7b3e963d09c2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.692790] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1331.692790] env[61995]: value = "task-795607" [ 1331.692790] env[61995]: _type = "Task" [ 1331.692790] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.699694] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795607, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.821595] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1331.821800] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1331.821983] env[61995]: DEBUG nova.network.neutron [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1332.203017] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795607, 'name': Rename_Task, 'duration_secs': 0.137727} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.203425] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1332.203559] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c82f596-bf23-44bd-b1de-d887bbd7d37e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.210983] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1332.210983] env[61995]: value = "task-795608" [ 1332.210983] env[61995]: _type = "Task" [ 1332.210983] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.217575] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795608, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.497918] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Applying migration context for instance ab8be459-0fb7-4f8d-ba53-66a7d64b30fc as it has an incoming, in-progress migration edcaea7f-f25c-4f42-b053-8d4480db3721. Migration status is post-migrating {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1332.498649] env[61995]: INFO nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating resource usage from migration edcaea7f-f25c-4f42-b053-8d4480db3721 [ 1332.506615] env[61995]: DEBUG nova.network.neutron [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [{"id": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "address": "fa:16:3e:c5:69:4f", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcacfdcc7-4c", "ovs_interfaceid": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.516903] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Migration edcaea7f-f25c-4f42-b053-8d4480db3721 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1332.517061] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance ab8be459-0fb7-4f8d-ba53-66a7d64b30fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.517186] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.517356] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1332.517492] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1332.563566] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13cfec8c-d863-4874-8c24-3b19c968cff6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.571436] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19964c2a-809c-4294-9530-16d763138df6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.602232] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004f1832-4b3b-4c9e-ab0c-85b7f2b3e31c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.609612] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d3ce61-504f-42e0-b02c-0baa51998b9b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.622571] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1332.719782] env[61995]: DEBUG oslo_vmware.api [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795608, 'name': PowerOnVM_Task, 'duration_secs': 0.435477} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.720063] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1332.720275] env[61995]: INFO nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Took 6.70 seconds to spawn the instance on the hypervisor. [ 1332.720458] env[61995]: DEBUG nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1332.721217] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408577ff-0d98-4168-9d7a-bd17d2f74b52 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.009403] env[61995]: DEBUG oslo_concurrency.lockutils [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1333.125204] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1333.240134] env[61995]: INFO nova.compute.manager [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Took 11.38 seconds to build instance. [ 1333.529800] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ac08be-29fe-487a-92a0-9b328e94cf63 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.548886] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a9428d-d5ce-432f-b7c6-c877a1524029 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.555950] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance 'ab8be459-0fb7-4f8d-ba53-66a7d64b30fc' progress to 83 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1333.629595] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1333.629769] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.138s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1333.742232] env[61995]: DEBUG oslo_concurrency.lockutils [None req-3f934013-6aa5-46cc-ae32-b6de8e581fbd tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.885s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1334.062834] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1334.063171] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c700c3c-1866-458c-b488-80e898fe74c8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.071012] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1334.071012] env[61995]: value = "task-795609" [ 1334.071012] env[61995]: _type = "Task" [ 1334.071012] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.080820] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795609, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.176145] env[61995]: DEBUG nova.compute.manager [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1334.176394] env[61995]: DEBUG nova.compute.manager [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing instance network info cache due to event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1334.176645] env[61995]: DEBUG oslo_concurrency.lockutils [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1334.176821] env[61995]: DEBUG oslo_concurrency.lockutils [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1334.177013] env[61995]: DEBUG nova.network.neutron [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1334.580217] env[61995]: DEBUG oslo_vmware.api [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795609, 'name': PowerOnVM_Task, 'duration_secs': 0.397774} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.580605] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1334.580692] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-192aba58-1941-41ec-a3cf-23ff577f215b tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance 'ab8be459-0fb7-4f8d-ba53-66a7d64b30fc' progress to 100 {{(pid=61995) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1334.877042] env[61995]: DEBUG nova.network.neutron [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updated VIF entry in instance network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1334.877440] env[61995]: DEBUG nova.network.neutron [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1335.380902] env[61995]: DEBUG oslo_concurrency.lockutils [req-2d1568f0-7eec-4a39-b427-d0ea36a7007e req-c22f70dd-954e-4e62-9b71-7492a54d98fa service nova] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1337.000629] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.000629] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1337.000629] env[61995]: DEBUG nova.compute.manager [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Going to confirm migration 10 {{(pid=61995) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1337.536013] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1337.536283] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquired lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1337.536471] env[61995]: DEBUG nova.network.neutron [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1337.536662] env[61995]: DEBUG nova.objects.instance [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'info_cache' on Instance uuid ab8be459-0fb7-4f8d-ba53-66a7d64b30fc {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1338.739086] env[61995]: DEBUG nova.network.neutron [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [{"id": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "address": "fa:16:3e:c5:69:4f", "network": {"id": "1c73e8b4-7a06-4d0d-b03d-ef749cf8b204", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-434071043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f3d88cb0eb54bc1a852c5f43f3806d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcacfdcc7-4c", "ovs_interfaceid": "cacfdcc7-4c0e-4623-80f0-0221590ba560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1339.241915] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Releasing lock "refresh_cache-ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1339.242236] env[61995]: DEBUG nova.objects.instance [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lazy-loading 'migration_context' on Instance uuid ab8be459-0fb7-4f8d-ba53-66a7d64b30fc {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1339.745841] env[61995]: DEBUG nova.objects.base [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1339.747183] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba6b1c9-1a09-4002-8e52-f579b2409529 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.766454] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41a7344d-c4c0-48f2-be67-2b59d3a0c412 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.771375] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1339.771375] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]520445ea-9c19-0559-d401-2c6136675790" [ 1339.771375] env[61995]: _type = "Task" [ 1339.771375] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.779658] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520445ea-9c19-0559-d401-2c6136675790, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.282520] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]520445ea-9c19-0559-d401-2c6136675790, 'name': SearchDatastore_Task, 'duration_secs': 0.015949} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.282816] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.283068] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.842507] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dee5a1-eddb-4acf-abdb-069162411929 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.851547] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad75fd52-75c6-46e7-8f2d-81a4e9f6177a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.900083] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10db27ff-8eba-4cb2-a02b-c50200d2fbc7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.910793] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7def27a-54b4-463e-8c6e-60693af83ee8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.929307] env[61995]: DEBUG nova.compute.provider_tree [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1341.432418] env[61995]: DEBUG nova.scheduler.client.report [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1342.443735] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.160s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1342.998558] env[61995]: INFO nova.scheduler.client.report [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted allocation for migration edcaea7f-f25c-4f42-b053-8d4480db3721 [ 1343.504619] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.504s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1345.013458] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.013824] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1345.013954] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.014157] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1345.014333] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1345.016561] env[61995]: INFO nova.compute.manager [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Terminating instance [ 1345.018372] env[61995]: DEBUG nova.compute.manager [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1345.018576] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1345.019443] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85a4cb3-3593-4c06-903b-d6bc62dc012f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.027653] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1345.027868] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93b62865-d7c5-4559-9c71-6a389b7c0349 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.033560] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1345.033560] env[61995]: value = "task-795610" [ 1345.033560] env[61995]: _type = "Task" [ 1345.033560] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.041051] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.544087] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795610, 'name': PowerOffVM_Task, 'duration_secs': 0.2417} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.544368] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1345.544546] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1345.544790] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-539ce2e5-2200-4e27-805f-9ca7cd5502f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.615246] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1345.615512] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1345.615719] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleting the datastore file [datastore2] ab8be459-0fb7-4f8d-ba53-66a7d64b30fc {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1345.615987] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2a9ae02-b50d-438d-849d-2601202bc982 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.622423] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for the task: (returnval){ [ 1345.622423] env[61995]: value = "task-795612" [ 1345.622423] env[61995]: _type = "Task" [ 1345.622423] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.629672] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795612, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.132947] env[61995]: DEBUG oslo_vmware.api [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Task: {'id': task-795612, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132858} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.133334] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1346.133433] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1346.133644] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1346.133856] env[61995]: INFO nova.compute.manager [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1346.134160] env[61995]: DEBUG oslo.service.loopingcall [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1346.134392] env[61995]: DEBUG nova.compute.manager [-] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1346.134496] env[61995]: DEBUG nova.network.neutron [-] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1346.584488] env[61995]: DEBUG nova.compute.manager [req-fa3e7f9a-d009-4952-9d1f-c267a1407077 req-4cf33cdd-d00c-4629-985a-a0551db6d559 service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Received event network-vif-deleted-cacfdcc7-4c0e-4623-80f0-0221590ba560 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1346.584692] env[61995]: INFO nova.compute.manager [req-fa3e7f9a-d009-4952-9d1f-c267a1407077 req-4cf33cdd-d00c-4629-985a-a0551db6d559 service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Neutron deleted interface cacfdcc7-4c0e-4623-80f0-0221590ba560; detaching it from the instance and deleting it from the info cache [ 1346.584855] env[61995]: DEBUG nova.network.neutron [req-fa3e7f9a-d009-4952-9d1f-c267a1407077 req-4cf33cdd-d00c-4629-985a-a0551db6d559 service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1347.064551] env[61995]: DEBUG nova.network.neutron [-] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1347.086954] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-981db4db-b417-4717-a44b-11a876ece554 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.097975] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ea22d2-55a6-40f5-8b8a-aecc96ab136a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.120415] env[61995]: DEBUG nova.compute.manager [req-fa3e7f9a-d009-4952-9d1f-c267a1407077 req-4cf33cdd-d00c-4629-985a-a0551db6d559 service nova] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Detach interface failed, port_id=cacfdcc7-4c0e-4623-80f0-0221590ba560, reason: Instance ab8be459-0fb7-4f8d-ba53-66a7d64b30fc could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1347.567758] env[61995]: INFO nova.compute.manager [-] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Took 1.43 seconds to deallocate network for instance. [ 1348.075160] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1348.075483] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1348.075700] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1348.097174] env[61995]: INFO nova.scheduler.client.report [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Deleted allocations for instance ab8be459-0fb7-4f8d-ba53-66a7d64b30fc [ 1348.605570] env[61995]: DEBUG oslo_concurrency.lockutils [None req-c4cff5b8-e74d-4a4b-a040-da9ec43b68f9 tempest-DeleteServersTestJSON-336282580 tempest-DeleteServersTestJSON-336282580-project-member] Lock "ab8be459-0fb7-4f8d-ba53-66a7d64b30fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.592s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.874062] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.874433] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.874481] env[61995]: INFO nova.compute.manager [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Shelving [ 1373.381914] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1373.382199] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76b0cff8-a58a-4330-aef8-8a696e346e3d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.391311] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1373.391311] env[61995]: value = "task-795614" [ 1373.391311] env[61995]: _type = "Task" [ 1373.391311] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1373.399500] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795614, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1373.900971] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795614, 'name': PowerOffVM_Task, 'duration_secs': 0.179984} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1373.901370] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1373.901939] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ffd66d-d26b-4d60-9892-97a623a8e8d8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.920643] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6d88a8-14ad-4f3d-9e6b-cd4ab60cfb39 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.430599] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Creating Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1374.430910] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-58d21783-8bc7-4b7c-8e61-805c7cd83a4f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.437999] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1374.437999] env[61995]: value = "task-795615" [ 1374.437999] env[61995]: _type = "Task" [ 1374.437999] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.445854] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795615, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1374.948475] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795615, 'name': CreateSnapshot_Task, 'duration_secs': 0.451567} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1374.948947] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Created Snapshot of the VM instance {{(pid=61995) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1374.949678] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99053207-28e7-44d2-91dd-094580819f4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.467604] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Creating linked-clone VM from snapshot {{(pid=61995) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1375.467983] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-888d724a-14ee-4abf-868f-76c417ce13db {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.481297] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1375.481297] env[61995]: value = "task-795616" [ 1375.481297] env[61995]: _type = "Task" [ 1375.481297] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1375.490709] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795616, 'name': CloneVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.991539] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795616, 'name': CloneVM_Task} progress is 94%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1376.492957] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795616, 'name': CloneVM_Task, 'duration_secs': 1.014743} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1376.493293] env[61995]: INFO nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Created linked-clone VM from snapshot [ 1376.494107] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9addcaf5-65e6-459f-8b1c-123c11ff65f2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.501418] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Uploading image e8802d9b-afa9-4597-ac24-a3dabf375b3a {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1376.521926] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1376.521926] env[61995]: value = "vm-185471" [ 1376.521926] env[61995]: _type = "VirtualMachine" [ 1376.521926] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1376.522201] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6e0d809d-8c9c-49b9-a387-4bee091367dc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.529147] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease: (returnval){ [ 1376.529147] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52939c18-b819-ee48-648e-79f6ba68f98b" [ 1376.529147] env[61995]: _type = "HttpNfcLease" [ 1376.529147] env[61995]: } obtained for exporting VM: (result){ [ 1376.529147] env[61995]: value = "vm-185471" [ 1376.529147] env[61995]: _type = "VirtualMachine" [ 1376.529147] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1376.529432] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the lease: (returnval){ [ 1376.529432] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52939c18-b819-ee48-648e-79f6ba68f98b" [ 1376.529432] env[61995]: _type = "HttpNfcLease" [ 1376.529432] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1376.535418] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1376.535418] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52939c18-b819-ee48-648e-79f6ba68f98b" [ 1376.535418] env[61995]: _type = "HttpNfcLease" [ 1376.535418] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1377.037907] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1377.037907] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52939c18-b819-ee48-648e-79f6ba68f98b" [ 1377.037907] env[61995]: _type = "HttpNfcLease" [ 1377.037907] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1377.038345] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1377.038345] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52939c18-b819-ee48-648e-79f6ba68f98b" [ 1377.038345] env[61995]: _type = "HttpNfcLease" [ 1377.038345] env[61995]: }. {{(pid=61995) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1377.038978] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b713f4a-7d35-4538-95c4-3b3a287c58ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.046081] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc4d34-e701-cc99-d846-2b267cab55ed/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1377.046261] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc4d34-e701-cc99-d846-2b267cab55ed/disk-0.vmdk for reading. {{(pid=61995) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1377.134493] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a7b83b94-ab61-41c6-bed7-64d38296787d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.950597] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc4d34-e701-cc99-d846-2b267cab55ed/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1384.951565] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c658d9-9c7b-47fd-a0f9-1752efa2e820 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.957530] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc4d34-e701-cc99-d846-2b267cab55ed/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1384.957705] env[61995]: ERROR oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc4d34-e701-cc99-d846-2b267cab55ed/disk-0.vmdk due to incomplete transfer. [ 1384.957918] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8f100461-8959-41e2-aec1-75b1f6828325 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.964755] env[61995]: DEBUG oslo_vmware.rw_handles [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc4d34-e701-cc99-d846-2b267cab55ed/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1384.964956] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Uploaded image e8802d9b-afa9-4597-ac24-a3dabf375b3a to the Glance image server {{(pid=61995) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1384.967281] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Destroying the VM {{(pid=61995) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1384.967512] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2e9b3f9d-8d60-45f2-b17c-c668d83b5aba {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.972566] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1384.972566] env[61995]: value = "task-795618" [ 1384.972566] env[61995]: _type = "Task" [ 1384.972566] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1384.981476] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795618, 'name': Destroy_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.482237] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795618, 'name': Destroy_Task, 'duration_secs': 0.338164} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1385.482515] env[61995]: INFO nova.virt.vmwareapi.vm_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Destroyed the VM [ 1385.482760] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Deleting Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1385.483012] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d69dbdcf-d465-41eb-b174-88b7b8667ae9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.489532] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1385.489532] env[61995]: value = "task-795619" [ 1385.489532] env[61995]: _type = "Task" [ 1385.489532] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.496322] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795619, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.999498] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795619, 'name': RemoveSnapshot_Task, 'duration_secs': 0.322865} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1385.999825] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Deleted Snapshot of the VM instance {{(pid=61995) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1386.000069] env[61995]: DEBUG nova.compute.manager [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1386.000751] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ac9672-fcbb-4955-884b-e535f7abf07e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.513195] env[61995]: INFO nova.compute.manager [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Shelve offloading [ 1386.514859] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1386.515125] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d4bc787-93f7-4db9-9291-a8a2b80819ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.521770] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1386.521770] env[61995]: value = "task-795620" [ 1386.521770] env[61995]: _type = "Task" [ 1386.521770] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1386.529162] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1387.031650] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] VM already powered off {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1387.031919] env[61995]: DEBUG nova.compute.manager [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1387.032587] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f76000-ee2a-4d4d-ae3a-16e9a46899ea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.038036] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1387.038179] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1387.038376] env[61995]: DEBUG nova.network.neutron [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1387.740068] env[61995]: DEBUG nova.network.neutron [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1388.242944] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1388.453784] env[61995]: DEBUG nova.compute.manager [req-9267ef36-ee40-43d4-a915-be829a5f82eb req-64bba7ad-86fa-4c0e-bf2c-a8ad77763ea9 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-vif-unplugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1388.454035] env[61995]: DEBUG oslo_concurrency.lockutils [req-9267ef36-ee40-43d4-a915-be829a5f82eb req-64bba7ad-86fa-4c0e-bf2c-a8ad77763ea9 service nova] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1388.454294] env[61995]: DEBUG oslo_concurrency.lockutils [req-9267ef36-ee40-43d4-a915-be829a5f82eb req-64bba7ad-86fa-4c0e-bf2c-a8ad77763ea9 service nova] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1388.454424] env[61995]: DEBUG oslo_concurrency.lockutils [req-9267ef36-ee40-43d4-a915-be829a5f82eb req-64bba7ad-86fa-4c0e-bf2c-a8ad77763ea9 service nova] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1388.454596] env[61995]: DEBUG nova.compute.manager [req-9267ef36-ee40-43d4-a915-be829a5f82eb req-64bba7ad-86fa-4c0e-bf2c-a8ad77763ea9 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] No waiting events found dispatching network-vif-unplugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1388.454792] env[61995]: WARNING nova.compute.manager [req-9267ef36-ee40-43d4-a915-be829a5f82eb req-64bba7ad-86fa-4c0e-bf2c-a8ad77763ea9 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received unexpected event network-vif-unplugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 for instance with vm_state shelved and task_state shelving_offloading. [ 1388.542336] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1388.543243] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25213063-0be3-4824-a446-68e2489a019c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.552219] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1388.552581] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4090362-6c61-4bc7-9e10-360c95bf0a2f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.624217] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1388.624598] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1388.624915] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleting the datastore file [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1388.625289] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a215cc2-c0a7-4692-81a0-2f287bbc9a4b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.632939] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1388.632939] env[61995]: value = "task-795622" [ 1388.632939] env[61995]: _type = "Task" [ 1388.632939] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.644366] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795622, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.016092] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.016360] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.141925] env[61995]: DEBUG oslo_vmware.api [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795622, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157018} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1389.142219] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1389.142411] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1389.142586] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1389.167071] env[61995]: INFO nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted allocations for instance 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 [ 1389.521622] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.521622] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1389.521622] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1389.671238] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1389.671499] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1389.671726] env[61995]: DEBUG nova.objects.instance [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'resources' on Instance uuid 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1390.060557] env[61995]: INFO nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating ports in neutron [ 1390.084538] env[61995]: INFO nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating port 6ab1828b-41a9-448a-853b-3ceb09e93d80 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1390.174157] env[61995]: DEBUG nova.objects.instance [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'numa_topology' on Instance uuid 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1390.478104] env[61995]: DEBUG nova.compute.manager [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1390.478399] env[61995]: DEBUG nova.compute.manager [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing instance network info cache due to event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1390.478645] env[61995]: DEBUG oslo_concurrency.lockutils [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.478810] env[61995]: DEBUG oslo_concurrency.lockutils [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.478977] env[61995]: DEBUG nova.network.neutron [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1390.677175] env[61995]: DEBUG nova.objects.base [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Object Instance<7ccf816b-ffa8-4f31-b0d1-2833ed8895d0> lazy-loaded attributes: resources,numa_topology {{(pid=61995) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1390.691133] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b83cbb-3c0c-4369-96c5-623c354e3291 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.699292] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024f8335-6029-4ea7-9878-baf36e1a99d2 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.727868] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aff1650-0cc6-4c84-8a0d-6ddc44584e4a {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.734334] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2299cbf4-1634-4f62-8fc3-39c0e5903743 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.748157] env[61995]: DEBUG nova.compute.provider_tree [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1391.177176] env[61995]: DEBUG nova.network.neutron [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updated VIF entry in instance network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1391.177575] env[61995]: DEBUG nova.network.neutron [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap6ab1828b-41", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1391.267526] env[61995]: ERROR nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [req-222783af-bfc5-48e8-bbc0-f961ca6d1773] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-222783af-bfc5-48e8-bbc0-f961ca6d1773"}]} [ 1391.284065] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1391.297277] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1391.297504] env[61995]: DEBUG nova.compute.provider_tree [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1391.308572] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1391.325959] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1391.342050] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1873809-36b2-4aee-8270-e10014c9c1fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.350489] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136b20bc-7577-420c-8a04-451e15ca09be {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.379602] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3a4a77-74a0-4727-9969-7a630559f579 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.386706] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf0445c-f0df-4778-b286-ea43d4f523a9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.399453] env[61995]: DEBUG nova.compute.provider_tree [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1391.609801] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1391.680385] env[61995]: DEBUG oslo_concurrency.lockutils [req-6d8e6aad-a1de-4534-aa5f-64dd0e1616e0 req-104c8226-10f7-48b1-a7a7-68d08ef7aba7 service nova] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1391.680680] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1391.680777] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1391.680953] env[61995]: DEBUG nova.objects.instance [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lazy-loading 'info_cache' on Instance uuid 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1391.919009] env[61995]: ERROR nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [req-31b8c3fe-e591-477f-bc11-27d331ee388a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 5c086f4d-bc91-4e49-9831-bed8df133c15. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-31b8c3fe-e591-477f-bc11-27d331ee388a"}]} [ 1391.933584] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing inventories for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1391.945020] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating ProviderTree inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1391.945134] env[61995]: DEBUG nova.compute.provider_tree [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 66, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1391.954535] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing aggregate associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, aggregates: None {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1391.972546] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Refreshing trait associations for resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61995) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1391.995379] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27cff46-9d10-4565-8d40-61be3797c7ff {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.003756] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43995a5a-b7d2-4aee-b13b-7a0ebeea725c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.034890] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4c98e0-16ac-4d44-9b28-45dc5e44d929 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.046184] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1513ed-60e3-4df1-bf26-817e8855c624 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.060359] env[61995]: DEBUG nova.compute.provider_tree [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1392.504586] env[61995]: DEBUG nova.compute.manager [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-vif-plugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1392.504794] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1392.505032] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1392.505219] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1392.505398] env[61995]: DEBUG nova.compute.manager [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] No waiting events found dispatching network-vif-plugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1392.505568] env[61995]: WARNING nova.compute.manager [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received unexpected event network-vif-plugged-6ab1828b-41a9-448a-853b-3ceb09e93d80 for instance with vm_state shelved_offloaded and task_state unshelving. [ 1392.505732] env[61995]: DEBUG nova.compute.manager [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1392.505889] env[61995]: DEBUG nova.compute.manager [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing instance network info cache due to event network-changed-6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1392.506068] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1392.541192] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1392.591952] env[61995]: DEBUG nova.scheduler.client.report [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updated inventory for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with generation 171 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1392.592255] env[61995]: DEBUG nova.compute.provider_tree [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating resource provider 5c086f4d-bc91-4e49-9831-bed8df133c15 generation from 171 to 172 during operation: update_inventory {{(pid=61995) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1392.592444] env[61995]: DEBUG nova.compute.provider_tree [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Updating inventory in ProviderTree for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1393.097982] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.426s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1393.402589] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1393.606960] env[61995]: DEBUG oslo_concurrency.lockutils [None req-55078a7b-22e9-4461-b379-2d995856b8d3 tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.733s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1393.608241] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.067s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1393.608456] env[61995]: INFO nova.compute.manager [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Unshelving [ 1393.905632] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1393.905830] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1393.906131] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1393.906305] env[61995]: DEBUG nova.network.neutron [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Refreshing network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1393.907813] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.907989] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.908405] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.908584] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.908733] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.908880] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.909014] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1393.909171] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.412065] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.412442] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.412442] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1394.412577] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1394.413477] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c965cd1-2906-47e9-bd09-86d7b407b330 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.421631] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3420761-bb05-422c-8646-aef70fea5fe6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.435397] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a7323e-9847-4342-a93c-3ed978006ac8 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.441651] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195640f4-8c69-496a-8901-2a7e6495645f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.471191] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181646MB free_disk=67GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1394.471465] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.471562] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.612913] env[61995]: DEBUG nova.network.neutron [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updated VIF entry in instance network info cache for port 6ab1828b-41a9-448a-853b-3ceb09e93d80. {{(pid=61995) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1394.613326] env[61995]: DEBUG nova.network.neutron [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1394.633288] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1395.117240] env[61995]: DEBUG oslo_concurrency.lockutils [req-1f214d1f-c882-4bee-b397-a002646617f6 req-79c3a9c9-e052-4a18-ac40-45390f3c2264 service nova] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1395.993321] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1395.993716] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1395.993716] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1396.019388] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8bbb35-705f-4eaa-b028-6923ef78ed4e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.027196] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93a6c2a-ee87-4ee9-8d4a-a082ae8c46ec {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.055871] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2805e7d5-f7be-42d0-8b01-ece3a3f0588d {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.062378] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac83a20b-4ea8-44f7-958b-17e6bc179956 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.074697] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1396.577830] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1397.082843] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1397.083181] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.612s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1397.083378] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.450s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1397.083603] env[61995]: DEBUG nova.objects.instance [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'pci_requests' on Instance uuid 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1397.588624] env[61995]: DEBUG nova.objects.instance [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'numa_topology' on Instance uuid 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1398.094095] env[61995]: INFO nova.compute.claims [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1399.129585] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee33756b-db24-43e6-8b01-cc575169bb68 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.137251] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aedf12e-af96-4b79-b255-326ee8694f4c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.166629] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c95681-2806-4b31-9264-beac991929bf {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.173826] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15b5291-37a3-4171-8ddf-e445e9552368 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.185459] env[61995]: DEBUG nova.compute.provider_tree [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1399.689400] env[61995]: DEBUG nova.scheduler.client.report [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1400.197594] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.114s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1400.232132] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1400.232455] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1400.232745] env[61995]: DEBUG nova.network.neutron [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Building network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1400.949810] env[61995]: DEBUG nova.network.neutron [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1401.452918] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1401.479890] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-10T08:52:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='931786121d765deefa91d6ce350b7798',container_format='bare',created_at=2024-09-10T09:07:04Z,direct_url=,disk_format='vmdk',id=e8802d9b-afa9-4597-ac24-a3dabf375b3a,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1499856703-shelved',owner='49e78af0ad9340258211bf92e447021c',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-09-10T09:07:16Z,virtual_size=,visibility=), allow threads: False {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1401.480165] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1401.480332] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image limits 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1401.480526] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Flavor pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1401.480712] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Image pref 0:0:0 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1401.480873] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61995) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1401.481106] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1401.481279] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1401.481455] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Got 1 possible topologies {{(pid=61995) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1401.481623] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1401.481800] env[61995]: DEBUG nova.virt.hardware [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61995) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1401.482672] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afeb98f0-d65b-440c-b6a2-a3dcc6ecd1ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.490643] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a7dac7-96af-4cde-9e90-0479a31376fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.503760] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:7c:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '28d04eee-6dbb-491a-a999-b659c799679d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ab1828b-41a9-448a-853b-3ceb09e93d80', 'vif_model': 'vmxnet3'}] {{(pid=61995) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1401.511067] env[61995]: DEBUG oslo.service.loopingcall [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1401.511348] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Creating VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1401.511662] env[61995]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5369b482-775e-4b10-9002-4bcd63f930a5 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.530231] env[61995]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1401.530231] env[61995]: value = "task-795623" [ 1401.530231] env[61995]: _type = "Task" [ 1401.530231] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.537320] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795623, 'name': CreateVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.039980] env[61995]: DEBUG oslo_vmware.api [-] Task: {'id': task-795623, 'name': CreateVM_Task, 'duration_secs': 0.319014} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1402.041032] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Created VM on the ESX host {{(pid=61995) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1402.041032] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1402.041196] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.041588] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1402.041839] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f455bf2-c8df-45ba-b016-d45602cec4c9 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.046270] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1402.046270] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52255b45-f79d-7d4a-65af-9db7ce369a46" [ 1402.046270] env[61995]: _type = "Task" [ 1402.046270] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.054987] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52255b45-f79d-7d4a-65af-9db7ce369a46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.556334] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1402.556618] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Processing image e8802d9b-afa9-4597-ac24-a3dabf375b3a {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1402.556829] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a/e8802d9b-afa9-4597-ac24-a3dabf375b3a.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1402.556978] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a/e8802d9b-afa9-4597-ac24-a3dabf375b3a.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.557185] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1402.557474] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-005a9b5c-01ec-4230-91f0-ccdd6891d711 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.565103] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1402.565302] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61995) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1402.565987] env[61995]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d98cf5d0-a28b-427a-9a7f-5389eb3f2c96 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.570671] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1402.570671] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52615648-b9b4-ee27-a551-f982e23377b3" [ 1402.570671] env[61995]: _type = "Task" [ 1402.570671] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.577751] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': session[528529ae-087a-3c5d-4ede-650c452f051a]52615648-b9b4-ee27-a551-f982e23377b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.080683] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Preparing fetch location {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1403.080968] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Fetch image to [datastore2] OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6/OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6.vmdk {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1403.081178] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Downloading stream optimized image e8802d9b-afa9-4597-ac24-a3dabf375b3a to [datastore2] OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6/OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6.vmdk on the data store datastore2 as vApp {{(pid=61995) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1403.081360] env[61995]: DEBUG nova.virt.vmwareapi.images [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Downloading image file data e8802d9b-afa9-4597-ac24-a3dabf375b3a to the ESX as VM named 'OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6' {{(pid=61995) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1403.146280] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1403.146280] env[61995]: value = "resgroup-9" [ 1403.146280] env[61995]: _type = "ResourcePool" [ 1403.146280] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1403.146594] env[61995]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-675b481f-a029-48d7-8a36-8801e1d1837e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.166796] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease: (returnval){ [ 1403.166796] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52cb5e74-416c-de34-bb32-db2b182854ed" [ 1403.166796] env[61995]: _type = "HttpNfcLease" [ 1403.166796] env[61995]: } obtained for vApp import into resource pool (val){ [ 1403.166796] env[61995]: value = "resgroup-9" [ 1403.166796] env[61995]: _type = "ResourcePool" [ 1403.166796] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1403.167166] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the lease: (returnval){ [ 1403.167166] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52cb5e74-416c-de34-bb32-db2b182854ed" [ 1403.167166] env[61995]: _type = "HttpNfcLease" [ 1403.167166] env[61995]: } to be ready. {{(pid=61995) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1403.172714] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1403.172714] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52cb5e74-416c-de34-bb32-db2b182854ed" [ 1403.172714] env[61995]: _type = "HttpNfcLease" [ 1403.172714] env[61995]: } is initializing. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1403.674740] env[61995]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1403.674740] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52cb5e74-416c-de34-bb32-db2b182854ed" [ 1403.674740] env[61995]: _type = "HttpNfcLease" [ 1403.674740] env[61995]: } is ready. {{(pid=61995) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1403.675210] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1403.675210] env[61995]: value = "session[528529ae-087a-3c5d-4ede-650c452f051a]52cb5e74-416c-de34-bb32-db2b182854ed" [ 1403.675210] env[61995]: _type = "HttpNfcLease" [ 1403.675210] env[61995]: }. {{(pid=61995) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1403.675748] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f6303d-7225-48df-ab3d-882f60589294 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.682541] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526d6ad2-845a-27e4-f06d-ba289e460e5c/disk-0.vmdk from lease info. {{(pid=61995) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1403.682658] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526d6ad2-845a-27e4-f06d-ba289e460e5c/disk-0.vmdk. {{(pid=61995) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1403.745688] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d4716ca7-f393-417e-b202-17f8214cefea {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.791195] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Completed reading data from the image iterator. {{(pid=61995) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1404.791670] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526d6ad2-845a-27e4-f06d-ba289e460e5c/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1404.792649] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38b77ce-e72e-49f5-be95-0b96d1dbc71f {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.800022] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526d6ad2-845a-27e4-f06d-ba289e460e5c/disk-0.vmdk is in state: ready. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1404.800022] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526d6ad2-845a-27e4-f06d-ba289e460e5c/disk-0.vmdk. {{(pid=61995) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1404.800022] env[61995]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-51d6ddaf-e695-4349-ba8c-32a6ba53cd02 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.001683] env[61995]: DEBUG oslo_vmware.rw_handles [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526d6ad2-845a-27e4-f06d-ba289e460e5c/disk-0.vmdk. {{(pid=61995) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1405.001939] env[61995]: INFO nova.virt.vmwareapi.images [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Downloaded image file data e8802d9b-afa9-4597-ac24-a3dabf375b3a [ 1405.003354] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a118ce-f964-4057-ba20-1b620f232dcd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.017880] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f36635a-0370-498f-a2f9-8a3ed4df7f74 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.046169] env[61995]: INFO nova.virt.vmwareapi.images [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] The imported VM was unregistered [ 1405.048512] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Caching image {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1405.048743] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1405.048989] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb76a51b-852b-4840-8dc1-0da4d3103aa7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.067309] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Created directory with path [datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a {{(pid=61995) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1405.067493] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6/OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6.vmdk to [datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a/e8802d9b-afa9-4597-ac24-a3dabf375b3a.vmdk. {{(pid=61995) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1405.067727] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d319e08f-2866-41ed-988d-a0399fed5c7e {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.074098] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1405.074098] env[61995]: value = "task-795626" [ 1405.074098] env[61995]: _type = "Task" [ 1405.074098] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.080830] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795626, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.586197] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795626, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.085754] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795626, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.587372] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795626, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.088149] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795626, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.587231] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795626, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.328737} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.587594] env[61995]: INFO nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6/OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6.vmdk to [datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a/e8802d9b-afa9-4597-ac24-a3dabf375b3a.vmdk. [ 1407.587720] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Cleaning up location [datastore2] OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6 {{(pid=61995) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1407.587879] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_889cff1f-5d68-4b11-88b9-b2fd1d5148d6 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1407.588144] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-667e8cec-2965-4ad0-9cab-e1740de634f0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.593838] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1407.593838] env[61995]: value = "task-795627" [ 1407.593838] env[61995]: _type = "Task" [ 1407.593838] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.601013] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795627, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.103349] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795627, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.075168} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.103772] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1408.103772] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a/e8802d9b-afa9-4597-ac24-a3dabf375b3a.vmdk" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1408.104018] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a/e8802d9b-afa9-4597-ac24-a3dabf375b3a.vmdk to [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1408.104273] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b3600c7c-72dc-4db5-88cb-dd378a317461 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.111306] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1408.111306] env[61995]: value = "task-795628" [ 1408.111306] env[61995]: _type = "Task" [ 1408.111306] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.118275] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795628, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.623316] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795628, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.123693] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795628, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.625260] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795628, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.126188] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795628, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.624852] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795628, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.172212} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1410.625126] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e8802d9b-afa9-4597-ac24-a3dabf375b3a/e8802d9b-afa9-4597-ac24-a3dabf375b3a.vmdk to [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk {{(pid=61995) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1410.625866] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf00d3f-115c-4fde-a949-5dac77bddfc7 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.646617] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1410.646841] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9963b9bb-f750-485c-94a7-d7fa1674389b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.665407] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1410.665407] env[61995]: value = "task-795629" [ 1410.665407] env[61995]: _type = "Task" [ 1410.665407] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1410.672479] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795629, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.175625] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795629, 'name': ReconfigVM_Task, 'duration_secs': 0.285536} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.176027] env[61995]: DEBUG nova.virt.vmwareapi.volumeops [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0/7ccf816b-ffa8-4f31-b0d1-2833ed8895d0.vmdk or device None with type streamOptimized {{(pid=61995) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1411.176519] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a2919b3-1299-41a0-aada-4a4440078ed6 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.182841] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1411.182841] env[61995]: value = "task-795630" [ 1411.182841] env[61995]: _type = "Task" [ 1411.182841] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.189831] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795630, 'name': Rename_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.695326] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795630, 'name': Rename_Task, 'duration_secs': 0.140816} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.695326] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powering on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1411.695326] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65b18803-e68e-4664-80e3-68893394630c {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.700099] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1411.700099] env[61995]: value = "task-795631" [ 1411.700099] env[61995]: _type = "Task" [ 1411.700099] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.714868] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795631, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.210402] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795631, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.711055] env[61995]: DEBUG oslo_vmware.api [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795631, 'name': PowerOnVM_Task, 'duration_secs': 0.581654} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1412.711366] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powered on the VM {{(pid=61995) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1412.802922] env[61995]: DEBUG nova.compute.manager [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Checking state {{(pid=61995) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1412.803856] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33ce9d3-da9e-4e50-8064-36b0eba7569b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.320879] env[61995]: DEBUG oslo_concurrency.lockutils [None req-6db5f434-e195-415e-8fc3-0a6e0c8f0d5b tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.713s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1441.595362] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1441.595884] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1442.098460] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1442.098737] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1442.098848] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1442.099016] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61995) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1442.099902] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8ecd9d-c287-4c82-ae12-d0d9a2c8fc25 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.108105] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d70f3b2-388a-478a-a646-bd8c2919d6fc {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.121314] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0221228f-0cea-4ff1-94be-3e6fa50d45bd {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.127533] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca60a9f1-c7f3-4af6-9a13-21526e324586 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.155011] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181663MB free_disk=67GB free_vcpus=48 pci_devices=None {{(pid=61995) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1442.155153] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1442.155336] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1443.249970] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Instance 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61995) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1443.250301] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1443.250301] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61995) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1443.277072] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dbef8e-b383-46b6-8862-6a9256bb28eb {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.284335] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b85aeb-7634-4f4d-844d-0fe845606d85 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.313556] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7d9bce-8f4d-4863-b22d-6ab0a59619e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.320021] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b95666-ef52-4fb2-831d-44df8edd06ed {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.332865] env[61995]: DEBUG nova.compute.provider_tree [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1443.836094] env[61995]: DEBUG nova.scheduler.client.report [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1444.342159] env[61995]: DEBUG nova.compute.resource_tracker [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61995) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1444.342535] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.187s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1444.342636] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1444.342775] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Cleaning up deleted instances with incomplete migration {{(pid=61995) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1445.845104] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1445.845552] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1445.845552] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Starting heal instance info cache {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1445.845754] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Rebuilding the list of instances to heal {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1446.375071] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquiring lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1446.375244] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Acquired lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1446.375393] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Forcefully refreshing network info cache for instance {{(pid=61995) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1446.375546] env[61995]: DEBUG nova.objects.instance [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Lazy-loading 'info_cache' on Instance uuid 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1448.107163] env[61995]: DEBUG nova.network.neutron [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [{"id": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "address": "fa:16:3e:3f:7c:7e", "network": {"id": "3b5d535a-17d5-49a1-a469-751106814597", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-45274755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49e78af0ad9340258211bf92e447021c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "28d04eee-6dbb-491a-a999-b659c799679d", "external-id": "nsx-vlan-transportzone-501", "segmentation_id": 501, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ab1828b-41", "ovs_interfaceid": "6ab1828b-41a9-448a-853b-3ceb09e93d80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1448.492994] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.493294] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1448.493522] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.493720] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1448.493892] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1448.497548] env[61995]: INFO nova.compute.manager [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Terminating instance [ 1448.499632] env[61995]: DEBUG nova.compute.manager [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Start destroying the instance on the hypervisor. {{(pid=61995) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1448.499835] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Destroying instance {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1448.500756] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3801f068-76cc-4385-8669-a70ece7f6734 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.509092] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powering off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1448.509323] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-496deb7e-34dd-4ed2-be26-b585bc25fca1 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.516072] env[61995]: DEBUG oslo_vmware.api [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1448.516072] env[61995]: value = "task-795632" [ 1448.516072] env[61995]: _type = "Task" [ 1448.516072] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1448.523404] env[61995]: DEBUG oslo_vmware.api [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1448.610293] env[61995]: DEBUG oslo_concurrency.lockutils [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Releasing lock "refresh_cache-7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" {{(pid=61995) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1448.610523] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updated the network info_cache for instance {{(pid=61995) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1448.610667] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.610831] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.610977] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.611133] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.611278] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.611406] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61995) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1449.026970] env[61995]: DEBUG oslo_vmware.api [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795632, 'name': PowerOffVM_Task, 'duration_secs': 0.180529} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1449.027297] env[61995]: DEBUG nova.virt.vmwareapi.vm_util [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Powered off the VM {{(pid=61995) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1449.027477] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Unregistering the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1449.027718] env[61995]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14c05333-c9e0-4890-8d92-8c4eb27db611 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.143306] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Unregistered the VM {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1449.143582] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Deleting contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1449.143822] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleting the datastore file [datastore2] 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1449.144105] env[61995]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85b82669-67a7-418d-b4ca-f548087ffe8b {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.151255] env[61995]: DEBUG oslo_vmware.api [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for the task: (returnval){ [ 1449.151255] env[61995]: value = "task-795634" [ 1449.151255] env[61995]: _type = "Task" [ 1449.151255] env[61995]: } to complete. {{(pid=61995) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1449.159026] env[61995]: DEBUG oslo_vmware.api [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795634, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1449.660494] env[61995]: DEBUG oslo_vmware.api [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Task: {'id': task-795634, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13296} completed successfully. {{(pid=61995) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1449.660780] env[61995]: DEBUG nova.virt.vmwareapi.ds_util [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted the datastore file {{(pid=61995) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1449.660973] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Deleted contents of the VM from datastore datastore2 {{(pid=61995) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1449.661173] env[61995]: DEBUG nova.virt.vmwareapi.vmops [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Instance destroyed {{(pid=61995) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1449.661362] env[61995]: INFO nova.compute.manager [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1449.661610] env[61995]: DEBUG oslo.service.loopingcall [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61995) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1449.661813] env[61995]: DEBUG nova.compute.manager [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Deallocating network for instance {{(pid=61995) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1449.661896] env[61995]: DEBUG nova.network.neutron [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] deallocate_for_instance() {{(pid=61995) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1450.098636] env[61995]: DEBUG nova.compute.manager [req-7b60a3bc-b244-4b9d-bf78-c7f747c99778 req-f9c584a4-fe8f-43e1-9ad7-5ba0b524105b service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Received event network-vif-deleted-6ab1828b-41a9-448a-853b-3ceb09e93d80 {{(pid=61995) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1450.098780] env[61995]: INFO nova.compute.manager [req-7b60a3bc-b244-4b9d-bf78-c7f747c99778 req-f9c584a4-fe8f-43e1-9ad7-5ba0b524105b service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Neutron deleted interface 6ab1828b-41a9-448a-853b-3ceb09e93d80; detaching it from the instance and deleting it from the info cache [ 1450.098993] env[61995]: DEBUG nova.network.neutron [req-7b60a3bc-b244-4b9d-bf78-c7f747c99778 req-f9c584a4-fe8f-43e1-9ad7-5ba0b524105b service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1450.572993] env[61995]: DEBUG nova.network.neutron [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Updating instance_info_cache with network_info: [] {{(pid=61995) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1450.601557] env[61995]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6569e8fd-b875-49dc-83ac-35e64690d8e0 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.611539] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c765ac9d-a53f-4166-a83a-bf90cf04c780 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.635207] env[61995]: DEBUG nova.compute.manager [req-7b60a3bc-b244-4b9d-bf78-c7f747c99778 req-f9c584a4-fe8f-43e1-9ad7-5ba0b524105b service nova] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Detach interface failed, port_id=6ab1828b-41a9-448a-853b-3ceb09e93d80, reason: Instance 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 could not be found. {{(pid=61995) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1451.075788] env[61995]: INFO nova.compute.manager [-] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Took 1.41 seconds to deallocate network for instance. [ 1451.583290] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1451.583626] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1451.583843] env[61995]: DEBUG nova.objects.instance [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lazy-loading 'resources' on Instance uuid 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 {{(pid=61995) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1452.120382] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d829c3a6-ee04-4787-bd3a-b050a8ed6017 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.127753] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9821d5-82a4-4a0f-bd67-b30e9d064252 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.156369] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b50040-1499-4993-b427-94109e30dc71 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.163269] env[61995]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684f1617-0aa0-46bd-90dc-17e293b6a196 {{(pid=61995) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.175516] env[61995]: DEBUG nova.compute.provider_tree [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed in ProviderTree for provider: 5c086f4d-bc91-4e49-9831-bed8df133c15 {{(pid=61995) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1452.595189] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1452.679359] env[61995]: DEBUG nova.scheduler.client.report [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Inventory has not changed for provider 5c086f4d-bc91-4e49-9831-bed8df133c15 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 67, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61995) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1453.184880] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1453.207513] env[61995]: INFO nova.scheduler.client.report [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Deleted allocations for instance 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0 [ 1453.715326] env[61995]: DEBUG oslo_concurrency.lockutils [None req-94864924-c169-463c-8702-fd12d09259aa tempest-AttachVolumeShelveTestJSON-1555801033 tempest-AttachVolumeShelveTestJSON-1555801033-project-member] Lock "7ccf816b-ffa8-4f31-b0d1-2833ed8895d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.222s {{(pid=61995) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1455.098756] env[61995]: DEBUG oslo_service.periodic_task [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61995) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1455.099147] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] Cleaning up deleted instances {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1455.605185] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] There are 14 instances to clean {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1455.605452] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 7ccf816b-ffa8-4f31-b0d1-2833ed8895d0] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1456.108747] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: ab8be459-0fb7-4f8d-ba53-66a7d64b30fc] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1456.611696] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 670c50f0-fd9b-4eb5-85c0-6dad9056b16c] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1457.115059] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 95af4a7a-bfe9-4bad-bb61-1df3736f81dc] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1457.618504] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 47db0510-411f-415c-99a4-63ed1bcbcbc9] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1458.122104] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 87efc871-d14b-4b9f-98c0-87474651cbff] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1458.624970] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: a3e4c652-49dc-48a1-affc-090e2b92736a] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1459.128752] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: c99b0295-bf1f-4a25-af31-cd0e19b6b089] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1459.632204] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 24853688-b972-48e6-be7b-7f0cf634c60d] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1460.136441] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 954d0ada-a870-415e-bd50-c066d27b7026] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1460.640129] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: da5c39c9-f733-4452-9c7c-c92830682428] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1461.144278] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 1c2afe20-e28f-4553-a6e1-7f31ec0db64b] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1461.648377] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: 1142205a-bc9e-4a94-b201-2fc203f8f913] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1462.151458] env[61995]: DEBUG nova.compute.manager [None req-0397e281-1a87-4f72-83dd-c5d524782ea5 None None] [instance: b2aa6712-3284-4d97-ac98-ff8789fa8bca] Instance has had 0 of 5 cleanup attempts {{(pid=61995) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}}